)]}'
{"/PATCHSET_LEVEL":[{"author":{"_account_id":34860,"name":"Amit Uniyal","email":"auniyal@redhat.com","username":"auniyal"},"change_message_id":"930298392719e7d85da30839f6b7d0a87cc48b2f","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":2,"id":"7a795782_80d0cb72","updated":"2023-08-29 07:12:42.000000000","message":"nova says volume is attached to VM, but its not listed as yet, tried restarted vm as well.","commit_id":"26ddbb8fac8940fa7b866067c9e3cd87bd5376d7"},{"author":{"_account_id":34860,"name":"Amit Uniyal","email":"auniyal@redhat.com","username":"auniyal"},"change_message_id":"dc0036bc8828ad484148cb25c0ee4554fddba15a","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":2,"id":"27946284_0d2a4477","updated":"2023-08-28 12:35:34.000000000","message":"tested with below steps:\n\n1- stopped  devstack@c-vol\n\n2- detach vol from server\n\n3- login and verify instance, if volume is listed using lsblk\n   server gets detached from instance\n\n4- openstack volume list\n    volume stayed in detaching state for more then 5-10 mins\n    then as expected went back (i.e. roll back) to in-use\n    https://paste.opendev.org/show/bbGUHTnxADFiJlLF3UkU/\n    \n5- login and verify instance, if volume is listed using lsblk\n   vol never not get listed again (waited more then 30 mins)\n\nextra lsblk logs\nhttps://paste.opendev.org/show/bnKaa5C0KdmVr6qHeVEY/","commit_id":"26ddbb8fac8940fa7b866067c9e3cd87bd5376d7"},{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"3ea340ff4f970a512dcefafd181167cbb5ac0635","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":2,"id":"f60b3160_c80fbcf5","in_reply_to":"27946284_0d2a4477","updated":"2024-09-04 18:36:57.000000000","message":"This is expected in the scenario you did. There are a few steps that happen when detaching a volume from an instance:\n\n1. The volume is detached from the guest (in Nova)\n2. The volume is detached from the compute host (by calling Cinder API)\n3. The volume status is set to \"available\" (by calling Cinder API)\n\nIn your test, you stopped the Cinder service. So Nova succeeds at step 1, detaching the volume from the guest (which you see inside the guest) but it times out asking Cinder to detach the volume from the compute host. By default, python-cinderclient does not set a timeout for the call to Cinder:\nhttps://github.com/openstack/python-cinderclient/blob/c4ecade6b16c25541d8e241c83f6c214ec5b437f/cinderclient/client.py#L403\nso the client will wait for a response for an undefined amount of time. It might be worth checking your nova-compute log to confirm whether the cinderclient took 5-10 minutes to time out before Nova could move on to roll back the volume status to \"in-use\" and that there was not some other reason for the delay in the status rollback.\n\nGoing back to the detach steps, if Nova succeeds in detaching the volume from the guest, it will not reattach it if the Cinder API call fails. This is currently the intended way that it works, that a detach can have made partial progress and the user needs to retry the detach later when Cinder is healthy again, to complete the total detach flow.","commit_id":"26ddbb8fac8940fa7b866067c9e3cd87bd5376d7"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"e668a511fe342347ee47b97d32a6e628d876c2dd","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":4,"id":"9a621d68_95478c60","updated":"2025-03-26 12:39:38.000000000","message":"i dont know if we want ot add a fucntional repoducer and or release note for this bug over all this makes sense i think\n\nrebased to see if this still applies cleanly and passes.","commit_id":"3363e917aef4ed1a641016191b6b16e1ac0f77ea"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"662e83b19eacd316e064fed2246eb0c664a9334e","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":5,"id":"c3a881ab_68891c85","updated":"2025-04-03 17:49:07.000000000","message":"+1 i need to review this properly but over all it looks ok\nim a littel to tired to dig into this today but ill try and loop back to it after the ptg.","commit_id":"f4db396c49c2257ce75ce06833e0ddbd8a3c8372"},{"author":{"_account_id":20865,"name":"Gökhan","email":"gokhan.isik@tubitak.gov.tr","username":"gokhan.isik"},"change_message_id":"921e68b9c593a5417c79d7fedc0630c9de2aa34e","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":6,"id":"d82c73f6_bb58a75a","updated":"2026-01-27 08:03:56.000000000","message":"recheck","commit_id":"2dcd18c3b84b42227f501aa3d40706c1d0ec77fc"},{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"c64e250c8e73a4e9928a566a393a9b7673d2d4d5","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":6,"id":"66edf456_fa23b6e3","updated":"2025-12-05 01:44:33.000000000","message":"recheck nova-next guest kernel panic","commit_id":"2dcd18c3b84b42227f501aa3d40706c1d0ec77fc"}],"nova/virt/block_device.py":[{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"77406fcb06da638f67da637d53c4ae9ba826e439","unresolved":true,"context_lines":[{"line_number":516,"context_line":"                volume_api.detach(context.elevated(), volume_id, instance.uuid,"},{"line_number":517,"context_line":"                                  attachment_id)"},{"line_number":518,"context_line":"            except Exception:"},{"line_number":519,"context_line":"                with excutils.save_and_reraise_exception():"},{"line_number":520,"context_line":"                    LOG.exception("},{"line_number":521,"context_line":"                        f\u0027Volume detach failed for volume {volume_id} \u0027"},{"line_number":522,"context_line":"                        f\u0027and attachment {attachment_id}\u0027, instance\u003dinstance"}],"source_content_type":"text/x-python","patch_set":2,"id":"e5c25a30_a274b74b","line":519,"range":{"start_line":519,"start_character":0,"end_line":519,"end_character":1},"updated":"2023-10-31 21:10:40.000000000","message":"we dont need to do this any more we can just call raise\nafter\nvolume_api.roll_detaching(context, volume_id)","commit_id":"26ddbb8fac8940fa7b866067c9e3cd87bd5376d7"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"e668a511fe342347ee47b97d32a6e628d876c2dd","unresolved":false,"context_lines":[{"line_number":516,"context_line":"                volume_api.detach(context.elevated(), volume_id, instance.uuid,"},{"line_number":517,"context_line":"                                  attachment_id)"},{"line_number":518,"context_line":"            except Exception:"},{"line_number":519,"context_line":"                with excutils.save_and_reraise_exception():"},{"line_number":520,"context_line":"                    LOG.exception("},{"line_number":521,"context_line":"                        f\u0027Volume detach failed for volume {volume_id} \u0027"},{"line_number":522,"context_line":"                        f\u0027and attachment {attachment_id}\u0027, instance\u003dinstance"}],"source_content_type":"text/x-python","patch_set":2,"id":"6b9d20c5_fe527f05","line":519,"range":{"start_line":519,"start_character":0,"end_line":519,"end_character":1},"in_reply_to":"e5c25a30_a274b74b","updated":"2025-03-26 12:39:38.000000000","message":"Done","commit_id":"26ddbb8fac8940fa7b866067c9e3cd87bd5376d7"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"77406fcb06da638f67da637d53c4ae9ba826e439","unresolved":true,"context_lines":[{"line_number":521,"context_line":"                        f\u0027Volume detach failed for volume {volume_id} \u0027"},{"line_number":522,"context_line":"                        f\u0027and attachment {attachment_id}\u0027, instance\u003dinstance"},{"line_number":523,"context_line":"                    )"},{"line_number":524,"context_line":"                    volume_api.roll_detaching(context, volume_id)"},{"line_number":525,"context_line":"        else:"},{"line_number":526,"context_line":"            try:"},{"line_number":527,"context_line":"                volume_api.attachment_delete(context, self[\u0027attachment_id\u0027])"}],"source_content_type":"text/x-python","patch_set":2,"id":"ca1db8e4_4992aa4d","line":524,"updated":"2023-10-31 21:10:40.000000000","message":"this will likely also fail correct.\nif we got a 500 or per haps a 503 rather then a 500 \n\nit does not hurt i guess too try an rool back but if we got a 500 we are in an undefiend state so we cant assume any future operators on teh volume will actully work.","commit_id":"26ddbb8fac8940fa7b866067c9e3cd87bd5376d7"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"e668a511fe342347ee47b97d32a6e628d876c2dd","unresolved":false,"context_lines":[{"line_number":521,"context_line":"                        f\u0027Volume detach failed for volume {volume_id} \u0027"},{"line_number":522,"context_line":"                        f\u0027and attachment {attachment_id}\u0027, instance\u003dinstance"},{"line_number":523,"context_line":"                    )"},{"line_number":524,"context_line":"                    volume_api.roll_detaching(context, volume_id)"},{"line_number":525,"context_line":"        else:"},{"line_number":526,"context_line":"            try:"},{"line_number":527,"context_line":"                volume_api.attachment_delete(context, self[\u0027attachment_id\u0027])"}],"source_content_type":"text/x-python","patch_set":2,"id":"ff6a54d3_c41faad2","line":524,"in_reply_to":"ca1db8e4_4992aa4d","updated":"2025-03-26 12:39:38.000000000","message":"Acknowledged","commit_id":"26ddbb8fac8940fa7b866067c9e3cd87bd5376d7"}]}
