)]}'
{"/PATCHSET_LEVEL":[{"author":{"_account_id":14567,"name":"Rodrigo Barbieri","email":"rodrigo.barbieri2010@gmail.com","username":"ganso"},"change_message_id":"3eb0e218b55ab6bb25fc29fa21c57ac2e165b174","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":8,"id":"febc7d6a_65a108d7","updated":"2022-02-04 17:12:06.000000000","message":"recheck","commit_id":"4e7d7b7d4d828f4dc7caf4f453ea25b72c1a190c"},{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"7b1c21aa5ccbe93b5dcec4a8f4b340f38479f1f9","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":9,"id":"e51a60f5_7d80db34","updated":"2022-02-09 18:12:13.000000000","message":"Going to fix this real quick","commit_id":"95157314bddd7eacc325f3f47a046f907abd8a87"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"0a47158e8ea37e9afba401e3fac54a049085cbd6","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":9,"id":"8fdad2ac_fa2c58d6","updated":"2022-02-09 18:02:43.000000000","message":"trivial rebase.\nso i spoke to frickler and clarkb about this on the infra channel\n\nzuul and gerrit have different requirement with regards to what is mergeable.\n\nzuul will generte a merge commit when its testing assuming there is no merge conflict\nbut gerrit cannot merge this because its parent is not in the target tree.\n\nbasically this patch\u0027s parent was an older revision of the one that finally merged.\nthat is fine form a zuul point of view but not from gerrits.","commit_id":"95157314bddd7eacc325f3f47a046f907abd8a87"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"01df8b1851509e52dd3863c44e38610c0fd7ca67","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":9,"id":"8021e009_50f0d576","in_reply_to":"8fdad2ac_fa2c58d6","updated":"2022-02-09 18:03:10.000000000","message":"and fungi","commit_id":"95157314bddd7eacc325f3f47a046f907abd8a87"}],"doc/source/admin/networking.rst":[{"author":{"_account_id":8864,"name":"Artom Lifshitz","email":"notartom@gmail.com","username":"artom"},"change_message_id":"62c06577f6b7d3554c7e39abed985219de5f165c","unresolved":true,"context_lines":[{"line_number":221,"context_line":"   The functionality described below is currently only supported by the"},{"line_number":222,"context_line":"   libvirt/KVM driver."},{"line_number":223,"context_line":""},{"line_number":224,"context_line":"Today\u0027s high-end servers have more processors, and guests running on them often"},{"line_number":225,"context_line":"have an increasing number of vCPUs. In a single virtio-net queue, the scale of"},{"line_number":226,"context_line":"the protocol stack in a guest is restricted, as the network performance does"},{"line_number":227,"context_line":"not scale as the number of vCPUs increases. Guests cannot transmit or retrieve"},{"line_number":228,"context_line":"packets in parallel, as virtio-net has only one TX and RX queue. The solution"},{"line_number":229,"context_line":"to this issue is to enable virtio-net multiqueue, which can allow the guest"},{"line_number":230,"context_line":"instances to increase the total network throughput."},{"line_number":231,"context_line":""},{"line_number":232,"context_line":"Before enabling multiqueue, you should assess the advantages and disadvantages"},{"line_number":233,"context_line":"of doing so. Multiqueue virtio-net provides the greatest performance benefit"}],"source_content_type":"text/x-rst","patch_set":2,"id":"eaef4780_13b2e3f8","line":230,"range":{"start_line":224,"start_character":0,"end_line":230,"end_character":51},"updated":"2021-05-20 14:23:28.000000000","message":"Can we rewrite this to feel less like the intro paragraph to a spec? :) I\u0027m thinking along the lines of \"Virtual NICs provided by the virtio-net driver support the multiqueue feature. Multiqueue allows the number of receive and transmit queues to scale with CPU count.\"","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"34458f642ce1f7a31b17dac3dd7fc5db29fa8941","unresolved":true,"context_lines":[{"line_number":221,"context_line":"   The functionality described below is currently only supported by the"},{"line_number":222,"context_line":"   libvirt/KVM driver."},{"line_number":223,"context_line":""},{"line_number":224,"context_line":"Today\u0027s high-end servers have more processors, and guests running on them often"},{"line_number":225,"context_line":"have an increasing number of vCPUs. In a single virtio-net queue, the scale of"},{"line_number":226,"context_line":"the protocol stack in a guest is restricted, as the network performance does"},{"line_number":227,"context_line":"not scale as the number of vCPUs increases. Guests cannot transmit or retrieve"},{"line_number":228,"context_line":"packets in parallel, as virtio-net has only one TX and RX queue. The solution"},{"line_number":229,"context_line":"to this issue is to enable virtio-net multiqueue, which can allow the guest"},{"line_number":230,"context_line":"instances to increase the total network throughput."},{"line_number":231,"context_line":""},{"line_number":232,"context_line":"Before enabling multiqueue, you should assess the advantages and disadvantages"},{"line_number":233,"context_line":"of doing so. Multiqueue virtio-net provides the greatest performance benefit"}],"source_content_type":"text/x-rst","patch_set":2,"id":"4b3430d1_49c0e433","line":230,"range":{"start_line":224,"start_character":0,"end_line":230,"end_character":51},"in_reply_to":"1bc0ae80_ecb2edb4","updated":"2021-05-20 19:59:12.000000000","message":"spcificlaly using udev\ne.g. for a 4 core vms \n\n# cat /etc/udev/rules.d/50-ethtool.rules\nACTION\u003d\u003d\"add\", SUBSYSTEM\u003d\u003d\"net\", NAME\u003d\u003d\"eth0\", RUN+\u003d\"/sbin/ethtool -L eth0 combined 4\"\n\nwe shoudl test that before we document it but i think that would be more or less correct","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"2dad2507ac595e5bf0bed60dfead705ae0f042d4","unresolved":false,"context_lines":[{"line_number":221,"context_line":"   The functionality described below is currently only supported by the"},{"line_number":222,"context_line":"   libvirt/KVM driver."},{"line_number":223,"context_line":""},{"line_number":224,"context_line":"Today\u0027s high-end servers have more processors, and guests running on them often"},{"line_number":225,"context_line":"have an increasing number of vCPUs. In a single virtio-net queue, the scale of"},{"line_number":226,"context_line":"the protocol stack in a guest is restricted, as the network performance does"},{"line_number":227,"context_line":"not scale as the number of vCPUs increases. Guests cannot transmit or retrieve"},{"line_number":228,"context_line":"packets in parallel, as virtio-net has only one TX and RX queue. The solution"},{"line_number":229,"context_line":"to this issue is to enable virtio-net multiqueue, which can allow the guest"},{"line_number":230,"context_line":"instances to increase the total network throughput."},{"line_number":231,"context_line":""},{"line_number":232,"context_line":"Before enabling multiqueue, you should assess the advantages and disadvantages"},{"line_number":233,"context_line":"of doing so. Multiqueue virtio-net provides the greatest performance benefit"}],"source_content_type":"text/x-rst","patch_set":2,"id":"753c3d49_3bdc585b","line":230,"range":{"start_line":224,"start_character":0,"end_line":230,"end_character":51},"in_reply_to":"4b3430d1_49c0e433","updated":"2021-05-25 16:18:48.000000000","message":"Done","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"e1292f5e1ca8eab9866be77c874526e970bda0f2","unresolved":true,"context_lines":[{"line_number":221,"context_line":"   The functionality described below is currently only supported by the"},{"line_number":222,"context_line":"   libvirt/KVM driver."},{"line_number":223,"context_line":""},{"line_number":224,"context_line":"Today\u0027s high-end servers have more processors, and guests running on them often"},{"line_number":225,"context_line":"have an increasing number of vCPUs. In a single virtio-net queue, the scale of"},{"line_number":226,"context_line":"the protocol stack in a guest is restricted, as the network performance does"},{"line_number":227,"context_line":"not scale as the number of vCPUs increases. Guests cannot transmit or retrieve"},{"line_number":228,"context_line":"packets in parallel, as virtio-net has only one TX and RX queue. The solution"},{"line_number":229,"context_line":"to this issue is to enable virtio-net multiqueue, which can allow the guest"},{"line_number":230,"context_line":"instances to increase the total network throughput."},{"line_number":231,"context_line":""},{"line_number":232,"context_line":"Before enabling multiqueue, you should assess the advantages and disadvantages"},{"line_number":233,"context_line":"of doing so. Multiqueue virtio-net provides the greatest performance benefit"}],"source_content_type":"text/x-rst","patch_set":2,"id":"fa9f00e2_d8f2bfe7","line":230,"range":{"start_line":224,"start_character":0,"end_line":230,"end_character":51},"in_reply_to":"eaef4780_13b2e3f8","updated":"2021-05-20 19:50:04.000000000","message":"i then to agree but for all the documenation not just the paragrah.\n\nalso form the spec i think the most impoatn part to document was how too enable it in the guest\n\nethtool -L \u003cNIC\u003e combined #num_of_queues\n\nwhich you have not included at all.","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"7a683e05063e375454540920d1ef749be4d824e8","unresolved":true,"context_lines":[{"line_number":221,"context_line":"   The functionality described below is currently only supported by the"},{"line_number":222,"context_line":"   libvirt/KVM driver."},{"line_number":223,"context_line":""},{"line_number":224,"context_line":"Today\u0027s high-end servers have more processors, and guests running on them often"},{"line_number":225,"context_line":"have an increasing number of vCPUs. In a single virtio-net queue, the scale of"},{"line_number":226,"context_line":"the protocol stack in a guest is restricted, as the network performance does"},{"line_number":227,"context_line":"not scale as the number of vCPUs increases. Guests cannot transmit or retrieve"},{"line_number":228,"context_line":"packets in parallel, as virtio-net has only one TX and RX queue. The solution"},{"line_number":229,"context_line":"to this issue is to enable virtio-net multiqueue, which can allow the guest"},{"line_number":230,"context_line":"instances to increase the total network throughput."},{"line_number":231,"context_line":""},{"line_number":232,"context_line":"Before enabling multiqueue, you should assess the advantages and disadvantages"},{"line_number":233,"context_line":"of doing so. Multiqueue virtio-net provides the greatest performance benefit"}],"source_content_type":"text/x-rst","patch_set":2,"id":"1bc0ae80_ecb2edb4","line":230,"range":{"start_line":224,"start_character":0,"end_line":230,"end_character":51},"in_reply_to":"fa9f00e2_d8f2bfe7","updated":"2021-05-20 19:54:39.000000000","message":"instead of deriving this form the spec i think we should praobly base this on our downstream docs\n\nhttps://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/7/html/virtualization_tuning_and_optimization_guide/sect-virtualization_tuning_optimization_guide-networking-techniques#sect-Virtualization_Tuning_Optimization_Guide-Networking-Multi-queue_virtio-net\n\nthis might also be useful for the guest configuration\nhttps://access.redhat.com/solutions/2127401","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":8864,"name":"Artom Lifshitz","email":"notartom@gmail.com","username":"artom"},"change_message_id":"62c06577f6b7d3554c7e39abed985219de5f165c","unresolved":true,"context_lines":[{"line_number":244,"context_line":"benefit, it has some limitations and therefore should not be unconditionally"},{"line_number":245,"context_line":"enabled:"},{"line_number":246,"context_line":""},{"line_number":247,"context_line":"- Guest OS is limited to ~200 MSI vectors. Each NIC queue requires a MSI"},{"line_number":248,"context_line":"  vector, as well as any virtio device or assigned PCI device. Defining an"},{"line_number":249,"context_line":"  instance with multiple virtio NICs and vCPUs might lead to a possibility of"},{"line_number":250,"context_line":"  hitting the guest MSI limit."}],"source_content_type":"text/x-rst","patch_set":2,"id":"639aa862_71406d19","line":247,"range":{"start_line":247,"start_character":30,"end_line":247,"end_character":33},"updated":"2021-05-20 14:23:28.000000000","message":"This acronym comes out of nowhere, and I have no idea what it means, even with a quick Google.","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"2dad2507ac595e5bf0bed60dfead705ae0f042d4","unresolved":false,"context_lines":[{"line_number":244,"context_line":"benefit, it has some limitations and therefore should not be unconditionally"},{"line_number":245,"context_line":"enabled:"},{"line_number":246,"context_line":""},{"line_number":247,"context_line":"- Guest OS is limited to ~200 MSI vectors. Each NIC queue requires a MSI"},{"line_number":248,"context_line":"  vector, as well as any virtio device or assigned PCI device. Defining an"},{"line_number":249,"context_line":"  instance with multiple virtio NICs and vCPUs might lead to a possibility of"},{"line_number":250,"context_line":"  hitting the guest MSI limit."}],"source_content_type":"text/x-rst","patch_set":2,"id":"a66bd6ba_3a4e32f9","line":247,"range":{"start_line":247,"start_character":30,"end_line":247,"end_character":33},"in_reply_to":"08a1e1ae_976732d4","updated":"2021-05-25 16:18:48.000000000","message":"I just dropped this. Agreed it\u0027s too into the weeds to warrant inclusion here","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"e1292f5e1ca8eab9866be77c874526e970bda0f2","unresolved":true,"context_lines":[{"line_number":244,"context_line":"benefit, it has some limitations and therefore should not be unconditionally"},{"line_number":245,"context_line":"enabled:"},{"line_number":246,"context_line":""},{"line_number":247,"context_line":"- Guest OS is limited to ~200 MSI vectors. Each NIC queue requires a MSI"},{"line_number":248,"context_line":"  vector, as well as any virtio device or assigned PCI device. Defining an"},{"line_number":249,"context_line":"  instance with multiple virtio NICs and vCPUs might lead to a possibility of"},{"line_number":250,"context_line":"  hitting the guest MSI limit."}],"source_content_type":"text/x-rst","patch_set":2,"id":"08a1e1ae_976732d4","line":247,"range":{"start_line":247,"start_character":30,"end_line":247,"end_character":33},"in_reply_to":"639aa862_71406d19","updated":"2021-05-20 19:50:04.000000000","message":"msi is message signaled interupts\nhttps://en.wikipedia.org/wiki/Message_Signaled_Interrupts\n\nand an msi vector is basicaly a mmio region or register into which a flag is set to notify the\nguest operating system that there is an interrupt that need to be handled.\n\n\nalthough we might actually be using msi-x in the q35 case.\n\nim not sure how accurate the 200 figure is any more by the way its probably still the case with pc.\nfor q35 its likely differnt.","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"e1292f5e1ca8eab9866be77c874526e970bda0f2","unresolved":true,"context_lines":[{"line_number":248,"context_line":"  vector, as well as any virtio device or assigned PCI device. Defining an"},{"line_number":249,"context_line":"  instance with multiple virtio NICs and vCPUs might lead to a possibility of"},{"line_number":250,"context_line":"  hitting the guest MSI limit."},{"line_number":251,"context_line":"- virtio-net multiqueue works well for incoming traffic, but can occasionally"},{"line_number":252,"context_line":"  cause a performance degradation, for outgoing traffic."},{"line_number":253,"context_line":"- Enabling virtio-net multiqueue increases the total network throughput, but in"},{"line_number":254,"context_line":"  parallel it also increases the CPU consumption."},{"line_number":255,"context_line":"- Enabling virtio-net multiqueue in the host QEMU config, does not enable the"}],"source_content_type":"text/x-rst","patch_set":2,"id":"62eacb15_278f8ea5","line":252,"range":{"start_line":251,"start_character":0,"end_line":252,"end_character":56},"updated":"2021-05-20 19:50:04.000000000","message":"im not sure if this is true anymore when we added multi queue support it was pretty \nnew in qemu so this feels like its proably outdated.","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"2dad2507ac595e5bf0bed60dfead705ae0f042d4","unresolved":false,"context_lines":[{"line_number":248,"context_line":"  vector, as well as any virtio device or assigned PCI device. Defining an"},{"line_number":249,"context_line":"  instance with multiple virtio NICs and vCPUs might lead to a possibility of"},{"line_number":250,"context_line":"  hitting the guest MSI limit."},{"line_number":251,"context_line":"- virtio-net multiqueue works well for incoming traffic, but can occasionally"},{"line_number":252,"context_line":"  cause a performance degradation, for outgoing traffic."},{"line_number":253,"context_line":"- Enabling virtio-net multiqueue increases the total network throughput, but in"},{"line_number":254,"context_line":"  parallel it also increases the CPU consumption."},{"line_number":255,"context_line":"- Enabling virtio-net multiqueue in the host QEMU config, does not enable the"}],"source_content_type":"text/x-rst","patch_set":2,"id":"c6ce0020_f23e03a7","line":252,"range":{"start_line":251,"start_character":0,"end_line":252,"end_character":56},"in_reply_to":"62eacb15_278f8ea5","updated":"2021-05-25 16:18:48.000000000","message":"Okay, dropped.","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"e1292f5e1ca8eab9866be77c874526e970bda0f2","unresolved":true,"context_lines":[{"line_number":255,"context_line":"- Enabling virtio-net multiqueue in the host QEMU config, does not enable the"},{"line_number":256,"context_line":"  functionality in the guest OS. The guest OS administrator needs to manually"},{"line_number":257,"context_line":"  turn it on for each guest NIC that requires this feature, using ethtool."},{"line_number":258,"context_line":"- MSI vectors would still be consumed (wasted), if multiqueue was enabled in"},{"line_number":259,"context_line":"  the host, but has not been enabled in the guest OS by the administrator."},{"line_number":260,"context_line":"- In case the number of vNICs in a guest instance is proportional to the number"},{"line_number":261,"context_line":"  of vCPUs, enabling the multiqueue feature is less important."},{"line_number":262,"context_line":"- Each virtio-net queue consumes 64 KB of kernel memory for the vhost driver."}],"source_content_type":"text/x-rst","patch_set":2,"id":"08eabe20_ab4d67e1","line":259,"range":{"start_line":258,"start_character":0,"end_line":259,"end_character":74},"updated":"2021-05-20 19:50:04.000000000","message":"this is partly true althoguh i think in newer linux kerenls the virtio-net-pci\ndriver may be able to auto enabel the queues.","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"2dad2507ac595e5bf0bed60dfead705ae0f042d4","unresolved":false,"context_lines":[{"line_number":255,"context_line":"- Enabling virtio-net multiqueue in the host QEMU config, does not enable the"},{"line_number":256,"context_line":"  functionality in the guest OS. The guest OS administrator needs to manually"},{"line_number":257,"context_line":"  turn it on for each guest NIC that requires this feature, using ethtool."},{"line_number":258,"context_line":"- MSI vectors would still be consumed (wasted), if multiqueue was enabled in"},{"line_number":259,"context_line":"  the host, but has not been enabled in the guest OS by the administrator."},{"line_number":260,"context_line":"- In case the number of vNICs in a guest instance is proportional to the number"},{"line_number":261,"context_line":"  of vCPUs, enabling the multiqueue feature is less important."},{"line_number":262,"context_line":"- Each virtio-net queue consumes 64 KB of kernel memory for the vhost driver."}],"source_content_type":"text/x-rst","patch_set":2,"id":"4df621bf_ead3dd31","line":259,"range":{"start_line":258,"start_character":0,"end_line":259,"end_character":74},"in_reply_to":"08eabe20_ab4d67e1","updated":"2021-05-25 16:18:48.000000000","message":"Ack","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"e1292f5e1ca8eab9866be77c874526e970bda0f2","unresolved":true,"context_lines":[{"line_number":259,"context_line":"  the host, but has not been enabled in the guest OS by the administrator."},{"line_number":260,"context_line":"- In case the number of vNICs in a guest instance is proportional to the number"},{"line_number":261,"context_line":"  of vCPUs, enabling the multiqueue feature is less important."},{"line_number":262,"context_line":"- Each virtio-net queue consumes 64 KB of kernel memory for the vhost driver."},{"line_number":263,"context_line":""},{"line_number":264,"context_line":"Having considered these points, multiqueue can be enabled or disabled using"},{"line_number":265,"context_line":"either the :nova:extra-spec:`hw:vif_multiqueue_enabled` flavor extra spec or"}],"source_content_type":"text/x-rst","patch_set":2,"id":"c3638afb_c7f3e876","line":262,"range":{"start_line":262,"start_character":2,"end_line":262,"end_character":77},"updated":"2021-05-20 19:50:04.000000000","message":"given this can change over time im not sure we should doument this.\ni would also assume it will chagne if you change the virtio queue lenght config options.\ne.g. longer queuse will need more memroy.","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"2dad2507ac595e5bf0bed60dfead705ae0f042d4","unresolved":false,"context_lines":[{"line_number":259,"context_line":"  the host, but has not been enabled in the guest OS by the administrator."},{"line_number":260,"context_line":"- In case the number of vNICs in a guest instance is proportional to the number"},{"line_number":261,"context_line":"  of vCPUs, enabling the multiqueue feature is less important."},{"line_number":262,"context_line":"- Each virtio-net queue consumes 64 KB of kernel memory for the vhost driver."},{"line_number":263,"context_line":""},{"line_number":264,"context_line":"Having considered these points, multiqueue can be enabled or disabled using"},{"line_number":265,"context_line":"either the :nova:extra-spec:`hw:vif_multiqueue_enabled` flavor extra spec or"}],"source_content_type":"text/x-rst","patch_set":2,"id":"c4892ba7_73177c1a","line":262,"range":{"start_line":262,"start_character":2,"end_line":262,"end_character":77},"in_reply_to":"c3638afb_c7f3e876","updated":"2021-05-25 16:18:48.000000000","message":"Ack","commit_id":"37665de37384f85ce911cf22072b8e5e528c78f0"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"8fa4f8f04289a1051a63602c02fbf025f12e7500","unresolved":true,"context_lines":[{"line_number":37,"context_line":"   Nova will ignore PCI devices reported by the hypervisor if the address is"},{"line_number":38,"context_line":"   outside of these ranges."},{"line_number":39,"context_line":""},{"line_number":40,"context_line":""},{"line_number":41,"context_line":"NUMA Affinity"},{"line_number":42,"context_line":"-------------"},{"line_number":43,"context_line":""}],"source_content_type":"text/x-rst","patch_set":6,"id":"e4955cb7_cc8d62c0","line":40,"updated":"2021-09-01 15:33:49.000000000","message":"nit: unrelated","commit_id":"f8e2a7c846dd209ee2334a92edd9ba9ea85c6305"},{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"f252e75d454bfb69c2bbe6045fc19d34291ec489","unresolved":false,"context_lines":[{"line_number":37,"context_line":"   Nova will ignore PCI devices reported by the hypervisor if the address is"},{"line_number":38,"context_line":"   outside of these ranges."},{"line_number":39,"context_line":""},{"line_number":40,"context_line":""},{"line_number":41,"context_line":"NUMA Affinity"},{"line_number":42,"context_line":"-------------"},{"line_number":43,"context_line":""}],"source_content_type":"text/x-rst","patch_set":6,"id":"f9dfe960_b0bb9940","line":40,"in_reply_to":"e4955cb7_cc8d62c0","updated":"2021-09-02 10:43:03.000000000","message":"Also incorrect. Most docs written in recent times put two lines before section headers (i.e. h2) to visually separate them.","commit_id":"f8e2a7c846dd209ee2334a92edd9ba9ea85c6305"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"8fa4f8f04289a1051a63602c02fbf025f12e7500","unresolved":true,"context_lines":[{"line_number":210,"context_line":".. versionchanged:: 24.0.0 (Xena)"},{"line_number":211,"context_line":""},{"line_number":212,"context_line":"   Support for configuring multiqueue via the ``hw:vif_multiqueue_enabled``"},{"line_number":213,"context_line":"   flavor extra spec was introduced in the Xena (23.0.0) release."},{"line_number":214,"context_line":""},{"line_number":215,"context_line":".. important::"},{"line_number":216,"context_line":""}],"source_content_type":"text/x-rst","patch_set":6,"id":"3fded3c5_2330b503","line":213,"range":{"start_line":213,"start_character":48,"end_line":213,"end_character":55},"updated":"2021-09-01 15:33:49.000000000","message":"no, 24.0.0 (as you said above)","commit_id":"f8e2a7c846dd209ee2334a92edd9ba9ea85c6305"},{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"f252e75d454bfb69c2bbe6045fc19d34291ec489","unresolved":false,"context_lines":[{"line_number":210,"context_line":".. versionchanged:: 24.0.0 (Xena)"},{"line_number":211,"context_line":""},{"line_number":212,"context_line":"   Support for configuring multiqueue via the ``hw:vif_multiqueue_enabled``"},{"line_number":213,"context_line":"   flavor extra spec was introduced in the Xena (23.0.0) release."},{"line_number":214,"context_line":""},{"line_number":215,"context_line":".. important::"},{"line_number":216,"context_line":""}],"source_content_type":"text/x-rst","patch_set":6,"id":"a2625fd8_85f4c5c9","line":213,"range":{"start_line":213,"start_character":48,"end_line":213,"end_character":55},"in_reply_to":"3fded3c5_2330b503","updated":"2021-09-02 10:43:03.000000000","message":"Good spot","commit_id":"f8e2a7c846dd209ee2334a92edd9ba9ea85c6305"},{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"7b1c21aa5ccbe93b5dcec4a8f4b340f38479f1f9","unresolved":true,"context_lines":[{"line_number":206,"context_line":""},{"line_number":207,"context_line":".. versionadded:: 12.0.0 (Liberty)"},{"line_number":208,"context_line":""},{"line_number":209,"context_line":".. versionchanged:: 24.0.0 (Xena)"},{"line_number":210,"context_line":""},{"line_number":211,"context_line":"   Support for configuring multiqueue via the ``hw:vif_multiqueue_enabled``"},{"line_number":212,"context_line":"   flavor extra spec was introduced in the Xena (24.0.0) release."}],"source_content_type":"text/x-rst","patch_set":9,"id":"649e1c44_6f6a9d2c","line":209,"range":{"start_line":209,"start_character":20,"end_line":209,"end_character":33},"updated":"2022-02-09 18:12:13.000000000","message":"Yoga (25.0.0)","commit_id":"95157314bddd7eacc325f3f47a046f907abd8a87"},{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"7b1c21aa5ccbe93b5dcec4a8f4b340f38479f1f9","unresolved":true,"context_lines":[{"line_number":209,"context_line":".. versionchanged:: 24.0.0 (Xena)"},{"line_number":210,"context_line":""},{"line_number":211,"context_line":"   Support for configuring multiqueue via the ``hw:vif_multiqueue_enabled``"},{"line_number":212,"context_line":"   flavor extra spec was introduced in the Xena (24.0.0) release."},{"line_number":213,"context_line":""},{"line_number":214,"context_line":".. important::"},{"line_number":215,"context_line":""}],"source_content_type":"text/x-rst","patch_set":9,"id":"d6c42e0e_8b036827","line":212,"range":{"start_line":212,"start_character":43,"end_line":212,"end_character":56},"updated":"2022-02-09 18:12:13.000000000","message":"Yoga (25.0.0)","commit_id":"95157314bddd7eacc325f3f47a046f907abd8a87"}]}
