)]}'
{".zuul.yaml":[{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"06acb3c793e4ef3dc587679feacf873f133b8a54","unresolved":true,"context_lines":[{"line_number":126,"context_line":"      tempest_test_regex: (^tempest\\.api\\.compute\\.admin\\.(test_live_migration|test_migration))"},{"line_number":127,"context_line":"      # revert this when bug #1940425 is fixed in neutron"},{"line_number":128,"context_line":"      tempest_exclude_regex: (test_live_migration_with_trunk)"},{"line_number":129,"context_line":"      devstack_localrc: \u0026uec_image_vars"},{"line_number":130,"context_line":"          # We are using the split kernel/initramfs image by default in an"},{"line_number":131,"context_line":"          # effort to reduce the occurrence of guest kernel panics in the APIC"},{"line_number":132,"context_line":"          # timer handler. We will cover testing of the full disk image by"}],"source_content_type":"text/x-yaml","patch_set":1,"id":"494af8d3_214766a5","line":129,"updated":"2024-05-15 17:39:54.000000000","message":"ok so this will still be used on the Nova live migration job\nok i thought it was one of the policy jobs but this is fine.","commit_id":"7edef346a8f3ef8f4c10fe0ce8b48128ba174772"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"834c6039bde1bc6ee39077bee51cd9d42255dc5e","unresolved":true,"context_lines":[{"line_number":126,"context_line":"      tempest_test_regex: (^tempest\\.api\\.compute\\.admin\\.(test_live_migration|test_migration))"},{"line_number":127,"context_line":"      # revert this when bug #1940425 is fixed in neutron"},{"line_number":128,"context_line":"      tempest_exclude_regex: (test_live_migration_with_trunk)"},{"line_number":129,"context_line":"      devstack_localrc: \u0026uec_image_vars"},{"line_number":130,"context_line":"          # We are using the split kernel/initramfs image by default in an"},{"line_number":131,"context_line":"          # effort to reduce the occurrence of guest kernel panics in the APIC"},{"line_number":132,"context_line":"          # timer handler. We will cover testing of the full disk image by"}],"source_content_type":"text/x-yaml","patch_set":1,"id":"bf41dacd_9a04edd2","line":129,"in_reply_to":"494af8d3_214766a5","updated":"2024-05-15 17:58:16.000000000","message":"I just left it on the job it was defined in. I think having a job is good, doesn\u0027t matter all that much which one it is, but we can move it if you feel strongly.","commit_id":"7edef346a8f3ef8f4c10fe0ce8b48128ba174772"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"20cbf7e6404f6d180447f75764d52ef07e45cc9a","unresolved":false,"context_lines":[{"line_number":126,"context_line":"      tempest_test_regex: (^tempest\\.api\\.compute\\.admin\\.(test_live_migration|test_migration))"},{"line_number":127,"context_line":"      # revert this when bug #1940425 is fixed in neutron"},{"line_number":128,"context_line":"      tempest_exclude_regex: (test_live_migration_with_trunk)"},{"line_number":129,"context_line":"      devstack_localrc: \u0026uec_image_vars"},{"line_number":130,"context_line":"          # We are using the split kernel/initramfs image by default in an"},{"line_number":131,"context_line":"          # effort to reduce the occurrence of guest kernel panics in the APIC"},{"line_number":132,"context_line":"          # timer handler. We will cover testing of the full disk image by"}],"source_content_type":"text/x-yaml","patch_set":1,"id":"218f1a08_30fd7fc5","line":129,"in_reply_to":"bf41dacd_9a04edd2","updated":"2024-05-15 18:15:09.000000000","message":"nope this is just not the one o thought it was on im fine with having it on this.\nits not the only job the does live migration wiht local storage.\nwe also do it in the hybrid plug job so we have coverage of block migration with both image types.\n\ni would be more concerned if it was on the ceph job since live migration with ceph is not tested in any of the other jobs that im aware of.","commit_id":"7edef346a8f3ef8f4c10fe0ce8b48128ba174772"},{"author":{"_account_id":17685,"name":"Elod Illes","email":"elod.illes@est.tech","username":"elod.illes"},"change_message_id":"1e42bc1b563f010ae8f10e751f5e237f9c909782","unresolved":true,"context_lines":[{"line_number":114,"context_line":"      devstack_localrc: \u0026uec_image_vars"},{"line_number":115,"context_line":"          # We are using the split kernel/initramfs image by default in an"},{"line_number":116,"context_line":"          # effort to reduce the occurrence of guest kernel panics in the APIC"},{"line_number":117,"context_line":"          # timer handler. We will cover testing of the full disk image by"},{"line_number":118,"context_line":"          # using it in the nova-next job."},{"line_number":119,"context_line":"          CIRROS_VERSION: 0.6.2"},{"line_number":120,"context_line":"          DEFAULT_IMAGE_NAME: cirros-0.6.2-x86_64-uec"},{"line_number":121,"context_line":"          DEFAULT_IMAGE_FILE_NAME: cirros-0.6.2-x86_64-uec.tar.gz"}],"source_content_type":"text/x-yaml","patch_set":2,"id":"a4f954f6_1ab4a75b","side":"PARENT","line":118,"range":{"start_line":117,"start_character":27,"end_line":118,"end_character":42},"updated":"2024-07-12 06:58:09.000000000","message":"nit: this is not true anymore","commit_id":"84b0a481fe86701f2baa562a051fafa765ccd011"},{"author":{"_account_id":17685,"name":"Elod Illes","email":"elod.illes@est.tech","username":"elod.illes"},"change_message_id":"c62cec6ccb225945fb06a2b69ff9bbe8d72c7b85","unresolved":true,"context_lines":[{"line_number":272,"context_line":"              block_migrate_cinder_iscsi: false"},{"line_number":273,"context_line":"    post-run: playbooks/nova-live-migration/post-run.yaml"},{"line_number":274,"context_line":""},{"line_number":275,"context_line":"- job:"},{"line_number":276,"context_line":"    name: nova-lvm"},{"line_number":277,"context_line":"    parent: devstack-tempest"},{"line_number":278,"context_line":"    description: |"},{"line_number":279,"context_line":"      Run tempest compute API tests using LVM image backend. This only runs"},{"line_number":280,"context_line":"      against nova/virt/libvirt/* changes."},{"line_number":281,"context_line":"    # Copy irrelevant-files from nova-dsvm-multinode-base and then exclude"},{"line_number":282,"context_line":"    # anything that is not in nova/virt/libvirt/* or nova/privsep/*."},{"line_number":283,"context_line":"    irrelevant-files:"},{"line_number":284,"context_line":"      - ^(?!.zuul.yaml)(?!nova/virt/libvirt/)(?!nova/privsep/).*$"},{"line_number":285,"context_line":"      - ^api-.*$"},{"line_number":286,"context_line":"      - ^(test-|)requirements.txt$"},{"line_number":287,"context_line":"      - ^.*\\.rst$"},{"line_number":288,"context_line":"      - ^.git.*$"},{"line_number":289,"context_line":"      - ^doc/.*$"},{"line_number":290,"context_line":"      - ^nova/hacking/.*$"},{"line_number":291,"context_line":"      - ^nova/locale/.*$"},{"line_number":292,"context_line":"      - ^nova/tests/.*$"},{"line_number":293,"context_line":"      - ^nova/test.py$"},{"line_number":294,"context_line":"      - ^releasenotes/.*$"},{"line_number":295,"context_line":"      - ^setup.cfg$"},{"line_number":296,"context_line":"      - ^tools/.*$"},{"line_number":297,"context_line":"      - ^tox.ini$"},{"line_number":298,"context_line":"    vars:"},{"line_number":299,"context_line":"      # Increase the swap size to avoid OOM errors when running the tests."},{"line_number":300,"context_line":"      configure_swap_size: 8192"},{"line_number":301,"context_line":"      # We use the \"all\" environment for tempest_test_regex and"},{"line_number":302,"context_line":"      # tempest_exclude_regex."},{"line_number":303,"context_line":"      tox_envlist: all"},{"line_number":304,"context_line":"      # Only run compute API tests."},{"line_number":305,"context_line":"      tempest_test_regex: ^tempest\\.api\\.compute"},{"line_number":306,"context_line":"      # Skip slow tests. Also, skip some volume detach tests until bug#1998148"},{"line_number":307,"context_line":"      # is fixed."},{"line_number":308,"context_line":"      tempest_exclude_regex: (^tempest\\.(api\\.compute\\.(volumes\\.test_attach_volume\\.AttachVolumeTestJSON\\.test_attach_detach_volume|servers\\.(test_server_rescue\\.ServerStableDeviceRescueTest\\.test_stable_device_rescue_disk_virtio_with_volume_attached|test_server_rescue_negative\\.ServerRescueNegativeTestJSON\\.test_rescued_vm_detach_volume)))|.*\\[.*\\bslow\\b.*\\])"},{"line_number":309,"context_line":"      devstack_local_conf:"},{"line_number":310,"context_line":"        test-config:"},{"line_number":311,"context_line":"          $TEMPEST_CONFIG:"},{"line_number":312,"context_line":"            compute-feature-enabled:"},{"line_number":313,"context_line":"              # NOTE(mriedem): resize of non-volume-backed lvm instances does"},{"line_number":314,"context_line":"              # not yet work (bug 1831657)."},{"line_number":315,"context_line":"              resize: false"},{"line_number":316,"context_line":"              cold_migration: false"},{"line_number":317,"context_line":"      devstack_localrc:"},{"line_number":318,"context_line":"        NOVA_BACKEND: LVM"},{"line_number":319,"context_line":"        # Do not waste time clearing volumes."},{"line_number":320,"context_line":"        LVM_VOLUME_CLEAR: none"},{"line_number":321,"context_line":"        # Increase the size of the swift loopback device to accommodate RAW"},{"line_number":322,"context_line":"        # snapshots from the LV based instance disks."},{"line_number":323,"context_line":"        # See bug #1913451 for more details."},{"line_number":324,"context_line":"        SWIFT_LOOPBACK_DISK_SIZE: 24G"},{"line_number":325,"context_line":"        # As above, increase the total image limit per tenant to 10G"},{"line_number":326,"context_line":"        GLANCE_LIMIT_IMAGE_SIZE_TOTAL: 10240"},{"line_number":327,"context_line":"      devstack_services:"},{"line_number":328,"context_line":"        # Disable non-essential services that we don\u0027t need for this job."},{"line_number":329,"context_line":"        c-bak: false"},{"line_number":330,"context_line":"        openstack-cli-server: true"},{"line_number":331,"context_line":""},{"line_number":332,"context_line":"- job:"},{"line_number":333,"context_line":"    name: nova-emulation"},{"line_number":334,"context_line":"    parent: devstack-tempest"}],"source_content_type":"text/x-yaml","patch_set":2,"id":"fe47f0e4_e1c1ccdb","line":331,"range":{"start_line":275,"start_character":2,"end_line":331,"end_character":0},"in_reply_to":"a260ca45_353444a5","updated":"2024-07-12 07:00:35.000000000","message":"These warnings are quite distracting :/ hence i\u0027ve created a patch the factor these out [1]. please review it if you want to get rid of these.\n\n[1] https://review.opendev.org/c/openstack/nova/+/922212","commit_id":"eed3e2b47ffea24d08ad7a85a4e9c36ef56d815e"}],"/PATCHSET_LEVEL":[{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"aa90c42749f21ed93514add6657e2591b5d0c6d1","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"03bd7bb1_955cdb1f","updated":"2024-05-20 15:17:43.000000000","message":"Looks good to me. We had intended use of the UEC image on all jobs to be a temporary thing because it seemed to help improve the horrendous-at-the-time pass rate on CI jobs last cycle, so it would be good to see how things are doing now with the full image.","commit_id":"7edef346a8f3ef8f4c10fe0ce8b48128ba174772"},{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"ff0258fe1105d58787a83173f4e68d5ad7f8e994","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"4a8d7ebe_fe5f7cab","updated":"2024-05-20 15:31:57.000000000","message":"This is just a rebase after fixing a typo in the patch underneath, so re-approving","commit_id":"eed3e2b47ffea24d08ad7a85a4e9c36ef56d815e"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"4ef97e40f2129a15b3784bc4771d25f6f7c86d67","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"131b3a38_81db076b","updated":"2024-05-21 16:08:23.000000000","message":"recheck cinder test failed to attach/detach","commit_id":"eed3e2b47ffea24d08ad7a85a4e9c36ef56d815e"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"5c2c98fff5a902bd262b214184d3f53658d40467","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"5651ab20_f528da97","updated":"2024-05-21 18:07:31.000000000","message":"recheck nova-lvm failed with 500 from glance. Glance shows errors talking to swift.","commit_id":"eed3e2b47ffea24d08ad7a85a4e9c36ef56d815e"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"87f844a89009df7348e677af8410465505dd952f","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"4da74534_9dd99a43","updated":"2024-05-21 13:40:43.000000000","message":"recheck several fails in nova-lvm, not guest kernel crashes and not OOM. Seems like timeouts with other services perhaps.","commit_id":"eed3e2b47ffea24d08ad7a85a4e9c36ef56d815e"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"bc8e74e71b02bc9120b23d8ad83dd27ce5083ce1","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"9415f6df_390d7fd1","updated":"2024-05-20 20:18:19.000000000","message":"recheck ssh timeout (no kernel crash)","commit_id":"eed3e2b47ffea24d08ad7a85a4e9c36ef56d815e"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"e8d5a67cb6856e5c4c88ae59fea6f040c510eb4e","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"8a3726fb_562cd6bc","updated":"2024-05-21 09:56:34.000000000","message":"recheck test_instances_with_cinder_volumes_on_all_compute_nodes fail with a device/resouce busy error when mounting, this is unrelated","commit_id":"eed3e2b47ffea24d08ad7a85a4e9c36ef56d815e"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"169a18b431ff1fca19331e1c2dde59138f51c3cc","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"a79dda2a_14585153","updated":"2024-05-20 22:34:48.000000000","message":"recheck timeout in pre-upgrade grenade job","commit_id":"eed3e2b47ffea24d08ad7a85a4e9c36ef56d815e"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"ce9f5191c676c067c7b709da920a0b3f09ad2141","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"8deaf64d_9711c706","updated":"2024-05-22 14:21:01.000000000","message":"recheck unrelated post failure on ovs job","commit_id":"eed3e2b47ffea24d08ad7a85a4e9c36ef56d815e"}]}
