)]}'
{".zuul.yaml":[{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"1f567fbff3bfe6818f45f8ed65461b7d748296d6","unresolved":true,"context_lines":[{"line_number":971,"context_line":"        # - nova-tox-functional-py39"},{"line_number":972,"context_line":"        # - nova-tox-functional-py310"},{"line_number":973,"context_line":"        - nova-ironic-shards"},{"line_number":974,"context_line":"        - ironic-tempest-ipa-wholedisk-direct-tinyipa-multinode-shard"},{"line_number":975,"context_line":"        # - tempest-integrated-compute:"},{"line_number":976,"context_line":"        #    # NOTE(gmann): Policies changes do not need to run all the"},{"line_number":977,"context_line":"        #    # integration test jobs. Running only tempest and grenade"}],"source_content_type":"text/x-yaml","patch_set":4,"id":"1e999e95_a87482b7","line":974,"updated":"2024-02-27 21:09:41.000000000","message":"with a minor change this works\nits not running a lot of test (4 that pass) \nso we likely will want to increase that but it does show that we can have multiple compute node nodes each with seperate shards and create nova instances that can be managed by each  compute service in turn.\n\n\ni am going to respin this patch based off of master to enable this job in the weekly and experimental pipeline\ni will then crate a follow up pr in nova ro ironic to add a post hook that will use osc to assert that only the ironic nodes we expect to be visible are\n\nseperatly form this Jay is working on a diffent singlel node job that will have some \"real\" nodes backed by vms and some fake nodes back bay the fake dirver.\n\nthat simpler job will show that we can have many ironic nodes not part of the \"nova\" shard and we can assert that nova only reporst and manages the nova ones.\n\nwe may decided to stay with that simpler singel node jobs in the mediaum to long term but i think https://zuul.opendev.org/t/openstack/build/2a3f86e4974b4a4a9e875477c77f81e0/logs already show that this si workign correctly\nand we can impvoe on it after feature freeze\n\n\nnot the imporant things to look out for are\n\non each node we have a diffent shard key\n\ncontoler has main-node\nhttps://zuul.opendev.org/t/openstack/build/2a3f86e4974b4a4a9e875477c77f81e0/log/controller/logs/screen-n-cpu.txt#385\ncompute has subnode\nhttps://zuul.opendev.org/t/openstack/build/2a3f86e4974b4a4a9e875477c77f81e0/log/compute1/logs/screen-n-cpu.txt#385\n\neach compute service  when it queies ironic for the aviable nodes gets 3\n\nhttps://zuul.opendev.org/t/openstack/build/2a3f86e4974b4a4a9e875477c77f81e0/log/controller/logs/screen-n-cpu.txt#1886\n\nhttps://zuul.opendev.org/t/openstack/build/2a3f86e4974b4a4a9e875477c77f81e0/log/compute1/logs/screen-n-cpu.txt#1105\n^ that took a while for all 3 to be alaivble but it got there\n\n\nwe can see that the controller (np0036885046) create 3 compute node records\n\nNo compute node record for np0036885046:257efac9-7899-4196-85cd-e2bc87a00545\nNo compute node record for np0036885046:5e79015a-3f3f-44c1-901c-2ecb55690243\nNo compute node record for np0036885046:4bd4dc4b-2c92-442e-8b10-977ef9eb89ed \n\nfor the compute/subnode (np0036885047) it also create 3 compute node records\n\nNo compute node record for np0036885047:5e0a1f36-621e-4129-bbf7-81c71fb0594b\nNo compute node record for np0036885047:ac00ce8c-0776-48ac-b647-beb8270426b3\nNo compute node record for np0036885047:6aa651dc-62b1-465f-b344-89521419621a\n\n\nimportantly the uuid which is also the ironic node uuid are diffent between\n\nwe can see in the ironic conductor logs\n\nhttps://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_2a3/910333/4/check/ironic-tempest-ipa-wholedisk-direct-tinyipa-multinode-shard/2a3f86e/controller/logs/screen-ir-cond.txt\n\n Feb 27 16:11:14.097959 np0036885046 ironic-conductor[107317]: INFO ironic.conductor.cleaning [None req-ed9298c6-7d98-49b4-9643-fed6e9b5eab2 None None] Automated cleaning is disabled via configuration, node Node(allocation_id\u003dNone,automated_clean\u003dNone,bios_interface\u003d\u0027no-bios\u0027,boot_interface\u003d\u0027ipxe\u0027,boot_mode\u003dNone,chassis_id\u003d1,clean_step\u003d{},conductor_affinity\u003d1,conductor_group\u003d\u0027\u0027,console_enabled\u003dFalse,console_interface\u003d\u0027no-console\u0027,created_at\u003d2024-02-27T15:19:20Z,deploy_interface\u003d\u0027direct\u0027,deploy_step\u003d{},description\u003dNone,driver\u003d\u0027ipmi\u0027,driver_info\u003d{\u0027ipmi_address\u0027: \u002710.208.224.208\u0027, \u0027ipmi_username\u0027: \u0027admin\u0027, \u0027ipmi_password\u0027: \u0027password\u0027, \u0027ipmi_port\u0027: 6230},driver_internal_info\u003d{\u0027deploy_steps\u0027: None, \u0027last_power_state_change\u0027: \u00272024-02-27T16:10:57.497679\u0027, \u0027agent_version\u0027: \u00279.9.1.dev5\u0027, \u0027agent_last_heartbeat\u0027: \u00272024-02-27T16:08:05.569468\u0027, \u0027hardware_manager_version\u0027: {\u0027generic_hardware_manager\u0027: \u00271.2\u0027}, \u0027agent_cached_deploy_steps_refreshed\u0027: \u00272024-02-27T16:07:12.048714\u0027},extra\u003d{},fault\u003dNone,firmware_interface\u003d\u0027no-firmware\u0027,id\u003d1,inspect_interface\u003d\u0027no-inspect\u0027,inspection_finished_at\u003dNone,inspection_started_at\u003dNone,instance_info\u003d{},instance_uuid\u003dNone,last_error\u003dNone,lessee\u003dNone,maintenance\u003dFalse,maintenance_reason\u003dNone,management_interface\u003d\u0027ipmitool\u0027,name\u003d\u0027node-0\u0027,network_data\u003d{},network_interface\u003d\u0027neutron\u0027,owner\u003dNone,parent_node\u003dNone,power_interface\u003d\u0027ipmitool\u0027,power_state\u003d\u0027power off\u0027,properties\u003d{\u0027cpu_arch\u0027: \u0027x86_64\u0027, \u0027capabilities\u0027: \u0027boot_mode:uefi\u0027, \u0027vendor\u0027: \u0027unknown\u0027},protected\u003dFalse,protected_reason\u003dNone,provision_state\u003d\u0027available\u0027,provision_updated_at\u003d2024-02-27T16:11:14Z,raid_config\u003d{},raid_interface\u003d\u0027no-raid\u0027,rescue_interface\u003d\u0027no-rescue\u0027,reservation\u003d\u0027np0036885046\u0027,resource_class\u003d\u0027baremetal\u0027,retired\u003dFalse,retired_reason\u003dNone,secure_boot\u003dNone,service_step\u003d{},shard\u003d\u0027main-node\u0027,storage_interface\u003d\u0027noop\u0027,target_power_state\u003dNone,target_provision_state\u003dNone,target_raid_config\u003d{},traits\u003dTraitList,updated_at\u003d2024-02-27T16:11:14Z,uuid\u003d4bd4dc4b-2c92-442e-8b10-977ef9eb89ed,vendor_interface\u003d\u0027ipmitool\u0027) has been successfully moved to AVAILABLE state\n \n\nthat the shard keys are set properly \n\nso shard\u003d\u0027main-node\u0027 is set on 4bd4dc4b-2c92-442e-8b10-977ef9eb89ed which is on the contoler \n\nso while the job dose not run a lot of test it is showing that this feature is working","commit_id":"94224b73d44f985fcef6d1de982d6c749a541c11"}],"/PATCHSET_LEVEL":[{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"ae986305e3111f6b26d37c00c282d8279bb1c64b","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":5,"id":"9c6206e5_1d3641d8","updated":"2024-02-29 17:11:07.000000000","message":"Looks OK to me","commit_id":"2fa5cb6530a06d2d4fa882832c6b5fb27d096fbd"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"8125e590e7a6fdeda8cc8c7da89b02c8c6e1c1f2","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":5,"id":"606fbb5d_c76b4980","updated":"2024-02-29 16:12:13.000000000","message":"So, I think we can merge this today, I\u0027m a bit sad it would be a bit late, but OK.","commit_id":"2fa5cb6530a06d2d4fa882832c6b5fb27d096fbd"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"55792bca8026d7cda251ebea2de00766531101b9","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":5,"id":"0c98d018_8ec9d247","updated":"2024-02-28 20:34:03.000000000","message":"Sylvain we need to decide tomorrow if this meets our minimum testing requirement to porgress with merging the shard support. i think we really should try and supprot this in Caracal and work on improvign the testing next cycle.","commit_id":"2fa5cb6530a06d2d4fa882832c6b5fb27d096fbd"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"7fb01e2495f247a973f8b66a29eba804c9b3fb3b","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":5,"id":"50da8cd9_7f3a98ad","updated":"2024-02-27 21:47:52.000000000","message":"check experimental","commit_id":"2fa5cb6530a06d2d4fa882832c6b5fb27d096fbd"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"3679a67e9e5e568450d999ede2ccffd7892bcc6b","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":5,"id":"a728d988_9e3d18b9","updated":"2024-02-28 11:17:01.000000000","message":"check experimental","commit_id":"2fa5cb6530a06d2d4fa882832c6b5fb27d096fbd"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"e75e1816ae91d08a37c49b0db15c449233d6e809","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":5,"id":"b0459cd6_45ceeed4","updated":"2024-02-29 16:11:29.000000000","message":"check experimental","commit_id":"2fa5cb6530a06d2d4fa882832c6b5fb27d096fbd"},{"author":{"_account_id":10342,"name":"Jay Faulkner","display_name":"JayF","email":"jay@jvf.cc","username":"JayF","status":"youtube.com/@oss-gr / podcast.gr-oss.io"},"change_message_id":"3af4ebc864e8df300258da0b492ee7757d747e8b","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":5,"id":"9e0a086d_522ccab9","updated":"2024-02-28 15:22:46.000000000","message":"check experimental\n\nIronic gate was broken yesterday by ubuntu package changes, we have fixes merged now","commit_id":"2fa5cb6530a06d2d4fa882832c6b5fb27d096fbd"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"9416781ef3220fc282e5c83430459413925b19e6","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":5,"id":"945138cd_9c9b9475","updated":"2024-02-28 13:04:40.000000000","message":"https://zuul.opendev.org/t/openstack/build/184b8882e2ea4eefa129af571b7883b3 it passed as an fyi in the recheck\n\ni belive the dnsmasq issue was an issue ironic were aware of and fixed yesterday.","commit_id":"2fa5cb6530a06d2d4fa882832c6b5fb27d096fbd"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"02d05b218a6ca00856904b62feeed4c6cbd2a50c","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":5,"id":"352d4f3c_4c25a8a1","updated":"2024-02-29 18:35:24.000000000","message":"once the relevent patches are merged ill recheck this","commit_id":"2fa5cb6530a06d2d4fa882832c6b5fb27d096fbd"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"28b41b1edefac600f846d9fa9e1985c40d0e8376","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":5,"id":"c161f511_352456ab","updated":"2024-02-29 18:27:53.000000000","message":"this is an easy fix i just need to remove the depends on\n\nbut that means i have to wait for the ironic patch to merge\nso ill do that then","commit_id":"2fa5cb6530a06d2d4fa882832c6b5fb27d096fbd"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"02d05b218a6ca00856904b62feeed4c6cbd2a50c","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":5,"id":"26e2d116_365b6430","in_reply_to":"c161f511_352456ab","updated":"2024-02-29 18:35:24.000000000","message":"details here\n\nhttps://review.opendev.org/c/openstack/ironic/+/894460/20//COMMIT_MSG#12","commit_id":"2fa5cb6530a06d2d4fa882832c6b5fb27d096fbd"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"e113e8b6f8c1ad8ce9acc4d0666522fafd5bf253","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":6,"id":"79fbf545_728d703b","updated":"2024-03-01 00:21:01.000000000","message":"deps have now merged so reapproving","commit_id":"3b25b1e94c512a1c8d921c5a578b45e85aff2496"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"595daa74e7ff7f61a35b1049cde25046e03b62be","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":6,"id":"0502c6d3_5a33e834","updated":"2024-03-02 21:18:52.000000000","message":"recheck nova-lvm failsed on unrelated test","commit_id":"3b25b1e94c512a1c8d921c5a578b45e85aff2496"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"00698261c157073b2d6d892fb2787948bf84eb85","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":6,"id":"ec9b5bb9_26c29e6a","updated":"2024-03-01 10:18:50.000000000","message":"recheck rbac job timed out","commit_id":"3b25b1e94c512a1c8d921c5a578b45e85aff2496"}]}
