)]}'
{"/PATCHSET_LEVEL":[{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"820b87b0f1f5d9aa0e8f0a42cb1ae4d12e4d7dc2","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"f4bf26ea_7caf8ea5","updated":"2023-03-27 15:28:55.000000000","message":"Thanks Eric, you jinxed me on that one !","commit_id":"29de62bf3b3bf5eda8986bc94babf1c94d67bd4e"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"37b0eaec4c390b110479fb3da2f6fc1fbcc1eb67","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"ca245b3f_52a887f6","updated":"2023-03-27 15:25:10.000000000","message":"if CI happy I\u0027m happy","commit_id":"29de62bf3b3bf5eda8986bc94babf1c94d67bd4e"},{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"cc1647e469d8d71e88c84c0afe87f1b9b42d75c6","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"cdbc6929_086171ab","updated":"2023-03-29 00:20:07.000000000","message":"recheck Details: Fault: {\u0027code\u0027: 500, \u0027created\u0027: \u00272023-03-28T22:14:38Z\u0027, \u0027message\u0027: \u0027Unexpected vif_type\u003dbinding_failed\u0027}.","commit_id":"29de62bf3b3bf5eda8986bc94babf1c94d67bd4e"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"99e3115bc33c0ed6a17281ef16ee7ef14840199d","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"94837a83_0153f96f","updated":"2023-03-29 10:42:08.000000000","message":"recheck SSH connection failed when copying the disk on the cold migration for live-migration job :\n\n\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: Command: scp -r /opt/stack/data/nova/instances/222f7711-eff7-4200-8f8b-bc51cfe73500_resize/disk 10.209.64.102:/opt/stack/data/nova/instances/222f7711-eff7-4200-8f8b-bc51cfe73500/disk\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: Exit code: 1\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: Stdout: \u0027\u0027\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: Stderr: \u0027kex_exchange_identification: Connection closed by remote host\\r\\nConnection closed by 10.209.64.102 port 22\\r\\nlost connection\\n\u0027\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500] Traceback (most recent call last):\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]   File \"/opt/stack/nova/nova/compute/manager.py\", line 10716, in _error_out_instance_on_exception\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]     yield\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]   File \"/opt/stack/nova/nova/compute/manager.py\", line 5980, in _resize_instance\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]     disk_info \u003d self.driver.migrate_disk_and_power_off(\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]   File \"/opt/stack/nova/nova/virt/libvirt/driver.py\", line 11570, in migrate_disk_and_power_off\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]     with excutils.save_and_reraise_exception():\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]   File \"/usr/local/lib/python3.10/dist-packages/oslo_utils/excutils.py\", line 227, in __exit__\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]     self.force_reraise()\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]   File \"/usr/local/lib/python3.10/dist-packages/oslo_utils/excutils.py\", line 200, in force_reraise\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]     raise self.value\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]   File \"/opt/stack/nova/nova/virt/libvirt/driver.py\", line 11549, in migrate_disk_and_power_off\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]     libvirt_utils.copy_image(from_path, img_path, host\u003ddest,\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]   File \"/opt/stack/nova/nova/virt/libvirt/utils.py\", line 294, in copy_image\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]     remote_filesystem_driver.copy_file(src, dest,\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]   File \"/opt/stack/nova/nova/virt/libvirt/volume/remotefs.py\", line 104, in copy_file\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]     self.driver.copy_file(src, dst, on_execute\u003don_execute,\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]   File \"/opt/stack/nova/nova/virt/libvirt/volume/remotefs.py\", line 196, in copy_file\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]     processutils.execute(\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]   File \"/usr/local/lib/python3.10/dist-packages/oslo_concurrency/processutils.py\", line 438, in execute\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500]     raise ProcessExecutionError(exit_code\u003d_returncode,\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500] oslo_concurrency.processutils.ProcessExecutionError: Unexpected error while running command.\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500] Command: scp -r /opt/stack/data/nova/instances/222f7711-eff7-4200-8f8b-bc51cfe73500_resize/disk 10.209.64.102:/opt/stack/data/nova/instances/222f7711-eff7-4200-8f8b-bc51cfe73500/disk\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500] Exit code: 1\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500] Stdout: \u0027\u0027\nMar 29 08:46:28.732477 np0033599111 nova-compute[83506]: ERROR nova.compute.manager [instance: 222f7711-eff7-4200-8f8b-bc51cfe73500] Stderr: \u0027kex_exchange_identification: Connection closed by remote host\\r\\nConnection closed by 10.209.64.102 port 22\\r\\nlost connection\\n\u0027\nhttps://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_c7f/878693/1/check/nova-live-migration/c7fd6c0/controller/logs/screen-n-cpu.txt","commit_id":"29de62bf3b3bf5eda8986bc94babf1c94d67bd4e"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"cfa5cfdff9b9b20ee4665947409b976bf3996d10","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"a6870360_1982405f","updated":"2023-03-29 07:40:54.000000000","message":"recheck bug 1946339","commit_id":"29de62bf3b3bf5eda8986bc94babf1c94d67bd4e"},{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"3dd047b615b38d71923f525471264202447c4a87","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"258deea5_06c9019d","updated":"2023-03-28 21:11:43.000000000","message":"recheck nova.exception.DeviceDetachFailed: Device detach failed for vdb: Run out of retry while detaching device vdb with device alias virtio-disk1 from instance 8e7aac92-016e-4b43-a962-b710185c63c8 from the live domain config. Device is still attached to the guest.","commit_id":"29de62bf3b3bf5eda8986bc94babf1c94d67bd4e"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"6219977955244491892f7b622a297e2f46bfc109","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"c078ccf7_62366a2f","updated":"2023-03-28 08:37:20.000000000","message":"recheck port binding failed on grenade-multinode due to ovn agent gone :\n\nMar 27 22:00:16.992306 np0033583487 neutron-server[160892]: DEBUG neutron.plugins.ml2.managers [req-818b5ad7-e407-4986-b48f-78ee164a93ae req-487fc387-3883-4e40-b298-46190bdbe727 service neutron] Attempting to bind port 53b89f87-7b22-491a-aa44-520a23807f7d by drivers ovn,logger on host np0033583488 at level 0 using segments [{\u0027id\u0027: \u0027033cc161-bd45-42c7-8dc1-ba9a95fb1b38\u0027, \u0027network_type\u0027: \u0027geneve\u0027, \u0027physical_network\u0027: None, \u0027segmentation_id\u0027: 50872, \u0027network_id\u0027: \u0027c83cbf24-7dbe-4345-bc72-deb9c7b31358\u0027}] {{(pid\u003d160892) _bind_port_level /opt/stack/new/neutron/neutron/plugins/ml2/managers.py:835}}\nMar 27 22:00:16.996060 np0033583487 neutron-server[160892]: WARNING neutron.plugins.ml2.drivers.ovn.mech_driver.mech_driver [req-818b5ad7-e407-4986-b48f-78ee164a93ae req-487fc387-3883-4e40-b298-46190bdbe727 service neutron] Refusing to bind port 53b89f87-7b22-491a-aa44-520a23807f7d to dead agent:  \u003cneutron.plugins.ml2.drivers.ovn.agent.neutron_agent.ControllerGatewayAgent object at 0x7f6678fad360\u003e\n[...]\nMar 27 22:00:16.997885 np0033583487 neutron-server[160892]: ERROR neutron.plugins.ml2.managers [req-818b5ad7-e407-4986-b48f-78ee164a93ae req-487fc387-3883-4e40-b298-46190bdbe727 service neutron] Failed to bind port 53b89f87-7b22-491a-aa44-520a23807f7d on host np0033583488 for vnic_type normal using segments [{\u0027id\u0027: \u0027033cc161-bd45-42c7-8dc1-ba9a95fb1b38\u0027, \u0027network_type\u0027: \u0027geneve\u0027, \u0027physical_network\u0027: None, \u0027segmentation_id\u0027: 50872, \u0027network_id\u0027: \u0027c83cbf24-7dbe-4345-bc72-deb9c7b31358\u0027}]\nhttps://872de5c590dd926ff0db-30e72828a36544d0c7466f2989d78bfe.ssl.cf1.rackcdn.com/878693/1/gate/nova-grenade-multinode/30736fe/controller/logs/screen-q-svc.txt\n","commit_id":"29de62bf3b3bf5eda8986bc94babf1c94d67bd4e"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"25e0af0546347150b0603faccd7b7f8a2ed6c0b7","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"ac026bd8_1bec6fca","updated":"2023-03-28 18:01:08.000000000","message":"recheck same ovn agent gone on grenade-multinode :\n\n\nMar 28 11:39:33.718558 np0033589393 neutron-server[161938]: DEBUG neutron.plugins.ml2.managers [req-eb59baf2-0927-4a57-9042-bfbc492f0648 req-3a452b03-d0f0-4ea0-bafa-f50b254be177 service neutron] Attempting to bind port 9a6be7ed-127b-4289-aaff-5143952c5fe1 by drivers ovn,logger on host np0033589393 at level 0 using segments [{\u0027id\u0027: \u0027d3b349e0-e5ba-417c-868a-c4d2b189352d\u0027, \u0027network_type\u0027: \u0027geneve\u0027, \u0027physical_network\u0027: None, \u0027segmentation_id\u0027: 47348, \u0027network_id\u0027: \u0027add20c55-34f6-4a8e-81ab-ff3d2fba2362\u0027}] {{(pid\u003d161938) _bind_port_level /opt/stack/new/neutron/neutron/plugins/ml2/managers.py:835}}\nMar 28 11:39:33.721229 np0033589393 neutron-server[161938]: WARNING neutron.plugins.ml2.drivers.ovn.mech_driver.mech_driver [req-eb59baf2-0927-4a57-9042-bfbc492f0648 req-3a452b03-d0f0-4ea0-bafa-f50b254be177 service neutron] Refusing to bind port 9a6be7ed-127b-4289-aaff-5143952c5fe1 to dead agent:  \u003cneutron.plugins.ml2.drivers.ovn.agent.neutron_agent.ControllerGatewayAgent object at 0x7f5550ab1510\u003e\nhttps://8c14ddfe56056e4e35ff-0ca364809dbe3b9250075209fe309a8e.ssl.cf5.rackcdn.com/878693/1/gate/nova-grenade-multinode/3d396f9/controller/logs/screen-q-svc.txt\n\nOVN agent logs : https://8c14ddfe56056e4e35ff-0ca364809dbe3b9250075209fe309a8e.ssl.cf5.rackcdn.com/878693/1/gate/nova-grenade-multinode/3d396f9/controller/logs/screen-q-ovn-metadata-agent.txt","commit_id":"29de62bf3b3bf5eda8986bc94babf1c94d67bd4e"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"425befc34b1f0834b368c2769046789fd65ec8a3","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"4711143a_b1adde1d","updated":"2023-03-27 18:39:05.000000000","message":"recheck volume already gone when cleaning up the test on ceph-multistore\n\n\ntraceback-1: {{{\nTraceback (most recent call last):\n  File \"/opt/stack/tempest/tempest/lib/common/utils/test_utils.py\", line 87, in call_and_ignore_notfound_exc\n    return func(*args, **kwargs)\n  File \"/opt/stack/tempest/tempest/api/volume/base.py\", line 154, in _delete_volume_for_cleanup\n    BaseVolumeTest.delete_volume(volumes_client, volume_id)\n  File \"/opt/stack/tempest/tempest/api/volume/base.py\", line 196, in delete_volume\n    client.delete_volume(volume_id)\n  File \"/opt/stack/tempest/tempest/lib/services/volume/v3/volumes_client.py\", line 145, in delete_volume\n    resp, body \u003d self.delete(url)\n  File \"/opt/stack/tempest/tempest/lib/common/rest_client.py\", line 339, in delete\n    return self.request(\u0027DELETE\u0027, url, extra_headers, headers, body)\n  File \"/opt/stack/tempest/tempest/lib/services/volume/base_client.py\", line 39, in request\n    resp, resp_body \u003d super(BaseClient, self).request(\n  File \"/opt/stack/tempest/tempest/lib/common/rest_client.py\", line 742, in request\n    self._error_checker(resp, resp_body)\n  File \"/opt/stack/tempest/tempest/lib/common/rest_client.py\", line 857, in _error_checker\n    raise exceptions.BadRequest(resp_body, resp\u003dresp)\ntempest.lib.exceptions.BadRequest: Bad request\nDetails: {\u0027code\u0027: 400, \u0027message\u0027: \u0027Invalid volume: Volume status must be available or error or error_restoring or error_extending or error_managing and must not be migrating, attached, belong to a group, have snapshots, awaiting a transfer, or be disassociated from snapshots after volume transfer.\u0027}\n}}}\n\nTraceback (most recent call last):\n  File \"/opt/stack/tempest/tempest/lib/common/utils/test_utils.py\", line 87, in call_and_ignore_notfound_exc\n    return func(*args, **kwargs)\n  File \"/opt/stack/tempest/tempest/common/waiters.py\", line 146, in wait_for_server_termination\n    raise lib_exc.TimeoutException\ntempest.lib.exceptions.TimeoutException: Request timed out\nDetails: None\n","commit_id":"29de62bf3b3bf5eda8986bc94babf1c94d67bd4e"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"a0df15d8a4cbca6c527f9ce28a6c82458e11865a","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"64667e16_8db8a47f","updated":"2023-03-27 15:27:34.000000000","message":"yep we are using typing to capture the api contracts fo our functions and i agree we should not use implict optional and state it explictly.","commit_id":"29de62bf3b3bf5eda8986bc94babf1c94d67bd4e"}]}
