)]}'
{"/PATCHSET_LEVEL":[{"author":{"_account_id":17685,"name":"Elod Illes","email":"elod.illes@est.tech","username":"elod.illes"},"change_message_id":"e3510ef1279da437891df8a3b115edc18348cec5","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"cc057218_3d521413","updated":"2021-12-07 19:12:23.000000000","message":"The backport looks good to me: valid, merged in stable/xena (when it was still master branch), changes \u0026 conflicts are noted and resolved correctly.","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"adc6d661ab7e55a707fc6d2561b7ed33c25c7e3b","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"672f555a_c113157d","updated":"2022-06-14 00:46:00.000000000","message":"by the way the same or similar kernel panics happened in both gate and check \nthe guess tried to acces memeory it shoudl not.\ni have seen this when we ran out o fmemory in the guest too in the past\n\nif this happens again with the recheck we might want to see if we have increased the ram used for cirros to 128mb on satable wallay\n\ntaht is needed for the 0.5.2 image if i remember correctly.\ni dont think this this is in anyway related to this change.","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"},{"author":{"_account_id":17685,"name":"Elod Illes","email":"elod.illes@est.tech","username":"elod.illes"},"change_message_id":"af79c8ba4595c93f6b0a64b5a65335539cfff75d","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"3a186925_cadf608e","updated":"2022-03-02 10:03:43.000000000","message":"recheck","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"},{"author":{"_account_id":7634,"name":"Takashi Natsume","email":"takanattie@gmail.com","username":"natsumet"},"change_message_id":"d35115857826bb9ff8ab6a60918dc63a09968231","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"56b67f98_803fbe3f","updated":"2021-12-11 14:00:01.000000000","message":"recheck","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"},{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"6d5ea8b8329dd90ee76cea94565cfc8c507eb5f2","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"d1f54383_a7654f15","updated":"2022-01-06 18:22:32.000000000","message":"recheck","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"},{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"dcb7774978ffa03d2e9c3025b8d19de2bcc71c68","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"5b7dc2b0_f8d7ae35","updated":"2022-06-01 03:48:24.000000000","message":"recheck\n\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver [None req-512f1b0d-61ca-47c2-ae0a-d1f649692c10 tempest-TestVolumeSwap-513368351 tempest-TestVolumeSwap-513368351-project-admin] Failure rebasing volume /dev/sdb on vdb.: libvirt.libvirtError: invalid argument: disk \u0027vdb\u0027 not found in domain\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver Traceback (most recent call last):\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver   File \"/opt/stack/nova/nova/virt/libvirt/driver.py\", line 2121, in _swap_volume\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver     dev.copy(conf.to_xml(), reuse_ext\u003dTrue)\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver   File \"/opt/stack/nova/nova/virt/libvirt/guest.py\", line 743, in copy\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver     return self._guest._domain.blockCopy(self._disk, dest_xml, flags\u003dflags)\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver   File \"/usr/local/lib/python3.8/dist-packages/eventlet/tpool.py\", line 190, in doit\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver     result \u003d proxy_call(self._autowrap, f, *args, **kwargs)\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver   File \"/usr/local/lib/python3.8/dist-packages/eventlet/tpool.py\", line 148, in proxy_call\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver     rv \u003d execute(f, *args, **kwargs)\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver   File \"/usr/local/lib/python3.8/dist-packages/eventlet/tpool.py\", line 129, in execute\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver     six.reraise(c, e, tb)\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver   File \"/usr/local/lib/python3.8/dist-packages/six.py\", line 703, in reraise\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver     raise value\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver   File \"/usr/local/lib/python3.8/dist-packages/eventlet/tpool.py\", line 83, in tworker\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver     rv \u003d meth(*args, **kwargs)\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver   File \"/usr/lib/python3/dist-packages/libvirt.py\", line 833, in blockCopy\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver     if ret \u003d\u003d -1: raise libvirtError (\u0027virDomainBlockCopy() failed\u0027, dom\u003dself)\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver libvirt.libvirtError: invalid argument: disk \u0027vdb\u0027 not found in domain\nMay 25 18:04:57.810605 ubuntu-focal-rax-dfw-0029779365 nova-compute[53778]: ERROR nova.virt.libvirt.driver ","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"},{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"c69cfb260f53ec073e77a4cb21c24d75f19a48f7","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"40d62f2d_166d016e","updated":"2022-06-04 05:20:11.000000000","message":"recheck dependency was rechecked","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"},{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"6a44a39aedd10b1cbf80ef753babb7e02da8dbc0","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"8aa68059_ddbdb93a","updated":"2022-05-25 17:03:28.000000000","message":"recheck dependency was rechecked ","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"},{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"854052c88f60c5c6642b4eaa3f496cf027688603","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"4e23a990_6b7303db","updated":"2022-05-03 21:23:17.000000000","message":"recheck get fresh CI results","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"},{"author":{"_account_id":8556,"name":"Ghanshyam Maan","display_name":"Ghanshyam Maan","email":"gmaan.os14@gmail.com","username":"ghanshyam"},"change_message_id":"8ea3cb4b1894ef8c6872d19a93b04d53e848a22c","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"4a459cac_9b81d57c","updated":"2022-06-04 15:17:56.000000000","message":"recheck grenade job failure pip issue is fixed by devstack bakport 805008","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"8d2ce9a76a9293b3d9ef8a2fc4c84a6da8d04f1f","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"d93a4fd0_503ab2ce","updated":"2022-06-14 00:42:55.000000000","message":"recheck vm kernel panic\ni think this can happen if we start live migrating before the voluem is fully attached.\n\n[   13.615091] WARNING: kernel stack frame pointer at 0000000013ca4972 in kworker/u2:1:25 has bad value 000000008c595e97\n[   13.615396] unwind stack type:0 next_sp:00000000b6a73617 mask:0x6 graph_idx:0\n ...\n[   13.620227]  ? virtqueue_add_sgs+0x7e/0x90\n[   13.709530]  ? virtio_queue_rq+0x318/0x4b0 [virtio_blk]\n[   13.709852]  ? blk_mq_dispatch_rq_list+0x93/0x540\n[   13.710093]  ? elv_rb_del+0x24/0x30\n[   13.710296]  ? deadline_remove_request+0x4e/0xb0\n[   13.710542]  ? blk_mq_do_dispatch_sched+0x83/0x100\n[   13.710792]  ? blk_mq_sched_dispatch_requests+0xfd/0x170\n[   13.711060]  ? __blk_mq_run_hw_queue+0x5a/0x110\n[   13.711300]  ? __blk_mq_delay_run_hw_queue+0x107/0x160\n[   13.711560]  ? blk_mq_run_hw_queue+0x56/0x100\n[   13.711795]  ? blk_mq_sched_insert_requests+0x7a/0x100\n[   13.712056]  ? blk_mq_flush_plug_list+0x1e6/0x2a0\n[   13.712308]  ? __update_load_avg_se+0x11c/0x2e0\n[   13.712547]  ? __update_load_avg_se+0x11c/0x2e0\n[   13.712782]  ? __update_load_avg_se+0x11c/0x2e0\n[   13.713021]  ? update_load_avg+0x554/0x620\n[   13.713238]  ? dequeue_entity+0x105/0x430\n[   13.713451]  ? update_load_avg+0x82/0x620\n[   13.713689]  ? set_next_entity+0xb8/0x1f0\n[   13.713909]  ? pick_next_task_fair+0x4e8/0x710\n[   13.714141]  ? dequeue_task+0xc3/0x2d0\n[   13.714353]  ? __schedule+0x2a8/0x670\n[   13.714558]  ? schedule+0x33/0xa0\n[   13.714752]  ? worker_thread+0xbc/0x410\n[   13.714967]  ? kthread+0x121/0x140\n[   13.715162]  ? process_one_work+0x3f0/0x3f0\n[   13.719802]  ? kthread_park+0xb0/0xb0\n[   13.720034]  ? ret_from_fork+0x35/0x40\n[   13.720394] Modules linked in: ip_tables x_tables nls_utf8 nls_iso8859_1 nls_ascii isofs hid_generic usbhid hid virtio_rng virtio_gpu drm_kms_helper syscopyarea sysfillrect sysimgblt fb_sys_fops ttm drm virtio_scsi virtio_net net_failover failover virtio_input virtio_blk qemu_fw_cfg 9pnet_virtio 9pnet pcnet32 8139cp mii ne2k_pci 8390 e1000\n[   13.723064] ---[ end trace ed592ebde2635574 ]---\n[   13.723411] RIP: 0010:kfree+0x1fc/0x210\n[   13.723638] Code: e7 e8 b8 60 fd ff e9 20 ff ff ff 4d 89 e9 41 b8 01 00 00 00 48 89 d9 48 89 da 4c 89 e6 4c 89 f7 e8 89 ce ff ff e9 01 ff ff ff \u003c0f\u003e 0b 4c 8b 25 6b a0 37 01 e9 38 fe ff ff 66 0f 1f 44 00 00 66 66\n[   13.724392] RSP: 0018:ffffb649c0003de8 EFLAGS: 00000046\n[   13.724648] RAX: ffff98b3c2f62000 RBX: ffff98b3c2f62000 RCX: ffff98b3c2f62000\n[   13.724967] RDX: 000000000000051d RSI: ffff98b3c762f140 RDI: ffff98b3c2f62000\n[   13.725285] RBP: ffffb649c0003e08 R08: 0000000000000000 R09: 0000000000000000\n[   13.725624] R10: 0000000000000000 R11: 0000000000000000 R12: ffffdea7c00bd880\n[   13.725942] R13: ffffffff8d237a33 R14: ffff98b3c70021c0 R15: ffff98b3c2f621a0\n[   13.726265] FS:  0000000000000000(0000) GS:ffff98b3c7600000(0000) knlGS:0000000000000000\n[   13.726629] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033\n[   13.726884] CR2: 00000000013a07b8 CR3: 000000000621a000 CR4: 00000000000006f0\n[   13.727370] Kernel panic - not syncing: Fatal exception in interrupt\n[   13.728516] Kernel Offset: 0xbc00000 from 0xffffffff81000000 (relocation range: 0xffffffff80000000-0xffffffffbfffffff)\n[   13.729246] ---[ end Kernel panic - not syncing: Fatal exception in interrupt ]---","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"},{"author":{"_account_id":32966,"name":"jiang guodao","email":"jguodao@vmware.com","username":"jguodao"},"change_message_id":"fe1f0d5e382947ab7c173b830c3ade8f1fdb2d32","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"120fe8c0_2e12a23b","updated":"2022-01-13 09:18:03.000000000","message":"vmware-recheck-patch","commit_id":"665c053315439e1345aa131f4839945d662fb3f3"}]}
