)]}'
{"/PATCHSET_LEVEL":[{"author":{"_account_id":8064,"name":"Jake Yip","email":"jake.yip@ardc.edu.au","username":"jake"},"change_message_id":"adade04977c284d751d61b067a14eb83274fd787","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"535fdc4e_320b1136","updated":"2025-05-07 09:05:26.000000000","message":"-1 for visibility","commit_id":"007cc112a97bd97144322a65d5f3f682f35c9ec5"},{"author":{"_account_id":14394,"name":"Dale Smith","email":"dale@catalystcloud.nz","username":"dalees"},"change_message_id":"fc6d9ed53a84fab65582655956bcb8d3a7d704ea","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"4db4f0a6_4abd7347","updated":"2025-05-02 00:54:15.000000000","message":"Related (but not dependent) change: https://review.opendev.org/c/openstack/magnum-capi-helm/+/946599","commit_id":"007cc112a97bd97144322a65d5f3f682f35c9ec5"},{"author":{"_account_id":14394,"name":"Dale Smith","email":"dale@catalystcloud.nz","username":"dalees"},"change_message_id":"6ad2bba0eedc7eaa894b17aecfeeaa142157af45","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"f48cafc2_23864c8f","updated":"2025-05-07 21:34:34.000000000","message":"Split logging change out into https://review.opendev.org/c/openstack/magnum/+/949091","commit_id":"9cbbf9e0001b2ad68753446a4d772aea248b4f47"}],"magnum/service/periodic.py":[{"author":{"_account_id":8064,"name":"Jake Yip","email":"jake.yip@ardc.edu.au","username":"jake"},"change_message_id":"adade04977c284d751d61b067a14eb83274fd787","unresolved":true,"context_lines":[{"line_number":67,"context_line":"        self.cluster \u003d cluster"},{"line_number":68,"context_line":""},{"line_number":69,"context_line":"    def update_status(self):"},{"line_number":70,"context_line":"        LOG.debug(\"Updating status for cluster %s\", self.cluster.uuid)"},{"line_number":71,"context_line":"        # get the driver for the cluster"},{"line_number":72,"context_line":"        cdriver \u003d driver.Driver.get_driver_for_cluster(self.ctx, self.cluster)"},{"line_number":73,"context_line":"        # ask the driver to sync status"}],"source_content_type":"text/x-python","patch_set":1,"id":"82cc43e7_be7540a3","line":70,"range":{"start_line":70,"start_character":0,"end_line":70,"end_character":2},"updated":"2025-05-07 09:05:26.000000000","message":"I love this change, but I\u0027m wary of the impact.\n\nShould we separate the logging id-\u003euuid change into a different change, and include a reno? That can get merged pretty easily.","commit_id":"007cc112a97bd97144322a65d5f3f682f35c9ec5"},{"author":{"_account_id":14394,"name":"Dale Smith","email":"dale@catalystcloud.nz","username":"dalees"},"change_message_id":"58109f976ce29df519d4b021f95942aab58e0b5d","unresolved":true,"context_lines":[{"line_number":67,"context_line":"        self.cluster \u003d cluster"},{"line_number":68,"context_line":""},{"line_number":69,"context_line":"    def update_status(self):"},{"line_number":70,"context_line":"        LOG.debug(\"Updating status for cluster %s\", self.cluster.uuid)"},{"line_number":71,"context_line":"        # get the driver for the cluster"},{"line_number":72,"context_line":"        cdriver \u003d driver.Driver.get_driver_for_cluster(self.ctx, self.cluster)"},{"line_number":73,"context_line":"        # ask the driver to sync status"}],"source_content_type":"text/x-python","patch_set":1,"id":"11b37430_16763124","line":70,"range":{"start_line":70,"start_character":0,"end_line":70,"end_character":2},"in_reply_to":"071f5de5_07712f36","updated":"2025-06-24 08:58:06.000000000","message":"I will review `_COMPLETE` state and the commit message.\n\n`_COMPLETE` was already in the state list, but there are others that are helpful to include here.","commit_id":"007cc112a97bd97144322a65d5f3f682f35c9ec5"},{"author":{"_account_id":14394,"name":"Dale Smith","email":"dale@catalystcloud.nz","username":"dalees"},"change_message_id":"84134d23c911f1563d51b392dc692b1de807b4b0","unresolved":false,"context_lines":[{"line_number":67,"context_line":"        self.cluster \u003d cluster"},{"line_number":68,"context_line":""},{"line_number":69,"context_line":"    def update_status(self):"},{"line_number":70,"context_line":"        LOG.debug(\"Updating status for cluster %s\", self.cluster.uuid)"},{"line_number":71,"context_line":"        # get the driver for the cluster"},{"line_number":72,"context_line":"        cdriver \u003d driver.Driver.get_driver_for_cluster(self.ctx, self.cluster)"},{"line_number":73,"context_line":"        # ask the driver to sync status"}],"source_content_type":"text/x-python","patch_set":1,"id":"c5de5697_239bf7b5","line":70,"range":{"start_line":70,"start_character":0,"end_line":70,"end_character":2},"in_reply_to":"11b37430_16763124","updated":"2025-06-26 02:04:26.000000000","message":"Done","commit_id":"007cc112a97bd97144322a65d5f3f682f35c9ec5"},{"author":{"_account_id":14394,"name":"Dale Smith","email":"dale@catalystcloud.nz","username":"dalees"},"change_message_id":"6ad2bba0eedc7eaa894b17aecfeeaa142157af45","unresolved":true,"context_lines":[{"line_number":67,"context_line":"        self.cluster \u003d cluster"},{"line_number":68,"context_line":""},{"line_number":69,"context_line":"    def update_status(self):"},{"line_number":70,"context_line":"        LOG.debug(\"Updating status for cluster %s\", self.cluster.uuid)"},{"line_number":71,"context_line":"        # get the driver for the cluster"},{"line_number":72,"context_line":"        cdriver \u003d driver.Driver.get_driver_for_cluster(self.ctx, self.cluster)"},{"line_number":73,"context_line":"        # ask the driver to sync status"}],"source_content_type":"text/x-python","patch_set":1,"id":"071f5de5_07712f36","line":70,"range":{"start_line":70,"start_character":0,"end_line":70,"end_character":2},"in_reply_to":"82cc43e7_be7540a3","updated":"2025-05-07 21:34:34.000000000","message":"Yeah, the biggest change is that status is sync\u0027d for all clusters in `_COMPLETE` state.\n\nI have been running this in production for some time, and it\u0027s useful to surface errors that occur without Magnum changes (VM failure, autoscaling events in CAPI).\n\nWithout the _COMPLETE I also wonder if some cluster health may get stuck in an incomplete state if the periodic update changes from IN_PROGRESS to _COMPLETE before health stabilises.","commit_id":"007cc112a97bd97144322a65d5f3f682f35c9ec5"}]}
