)]}'
{"/PATCHSET_LEVEL":[{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"276cb0a652aee4c75303a21101413be2b4d943c9","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":16,"id":"3f83c9d2_39cd5e7e","updated":"2024-02-25 22:37:32.000000000","message":"this will not be required once the share mounting moves to the instance start.","commit_id":"afc7a2ca1e5b7b508c79c78ce537cf746feaef3d"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"699b87751765e9a91cc0b37566215e8153f4a6ef","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":26,"id":"737bfc11_324e3680","updated":"2024-07-02 11:25:06.000000000","message":"the mount share call is still missing from the _resume_guests_state code path.","commit_id":"f0be8d95f6d42fe06db5d21f951209618ffc3fb0"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"4c6f61868e06b1a1fa7d1da5261b7becd7c4dbdd","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":26,"id":"cd00e788_645f95b9","in_reply_to":"737bfc11_324e3680","updated":"2024-08-28 12:58:57.000000000","message":"I think we still missing the share mounting after host reboot during resume_state_on_host_boot","commit_id":"f0be8d95f6d42fe06db5d21f951209618ffc3fb0"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"9e4cc323af7bf73ea257330d02c274febd609139","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":26,"id":"233dd509_d19f3af8","in_reply_to":"cd00e788_645f95b9","updated":"2024-08-28 13:18:23.000000000","message":"OK it is added to the next commit in the series by mistake. As soon as that is change is pulled back to this I\u0027m OK.","commit_id":"f0be8d95f6d42fe06db5d21f951209618ffc3fb0"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"0b8c5a00bfe5e8849ec1b286485cbe0b15dbd713","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":27,"id":"e2abbffe_5293377b","updated":"2024-07-25 14:29:57.000000000","message":"I don\u0027t see the difference with the previous PS.","commit_id":"c733dadca577860a09de0e868059aaa9b8c7dbe1"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"2d45492b5e69b778a6f7d9f5ebab938be469aeef","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":31,"id":"a50925c0_d6b2773e","updated":"2024-09-17 12:12:36.000000000","message":"Thanks for adding the mount call. I tried to confirm that this works end to end in devstack but something is still not correct, but at this time I\u0027m not sure it is nova or manila.\n\nI did the following:\n* boot a VM, stop it, attach a share, start the VM\n* I mounted the share in the guest and wrote data to it, then unmounted the share.\n* I configured resume_guests_state_on_host_boot \u003d True in the nova-cpu.conf\n* I rebooted the devstack VM\nAfter the reboot I observed that nova did restarted the VM at the nova-compute startup and the share is in the domain XML. Also I confirmed that the share is in OK state in manila. I was able to mount the share in the rebooted guest, but the previously written content wasn\u0027t there. Could you please try to follow this reproduction to see if this is reproducible in your env as well or is it something specific to my env.","commit_id":"1214701997babaa8c75fd43cffb5eb97b118602a"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"57f7e466b94151c5d489fc426c126ccae8ab7733","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":31,"id":"31503c75_26899cb2","in_reply_to":"a50925c0_d6b2773e","updated":"2024-10-11 11:38:07.000000000","message":"This remains an open question.","commit_id":"1214701997babaa8c75fd43cffb5eb97b118602a"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"9c1e3b31d22241b1e45f337f15cac3604f0a8d24","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":34,"id":"ad6df19d_67a9cd18","updated":"2024-11-13 14:38:06.000000000","message":"-1 as the current patch doesn\u0027t work due to a missing token.","commit_id":"ece68a4df2d7ca206cbad74dc48405a3c520c1c7"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"3f41c0301d09e4106d18e1fceebf4d37044dfc3c","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":34,"id":"855adb4c_554d208c","updated":"2024-11-05 13:00:11.000000000","message":"Just rebase since my last review. We still need to manually test resume guest state and host reboot.","commit_id":"ece68a4df2d7ca206cbad74dc48405a3c520c1c7"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"a9af8eae41af738f47daae91597462cbb6ba842f","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":35,"id":"f50ad2be_1f98dd94","updated":"2024-11-15 10:24:33.000000000","message":"Looks good. Thanks @sbauza@redhat.com for tracking down the resume issue.","commit_id":"1e3143632f30065047ad325d84b93a545aaca4bd"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"87ec819b32c7fee2eba0ae4d1e0540c51e177650","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":35,"id":"1cdbcd5f_5dd2e597","updated":"2024-11-14 13:20:59.000000000","message":"the issue I said is now fixed, I tested it.\n\n```\nNov 14 09:59:30 sbauza-manila2 nova-compute[827]: INFO nova.compute.manager [None req-4bbbdf5d-d998-4ebf-b5a7-157eaeeef1e1 None None] [instance: 4b57bbae-ae14-47b5-a529-901753d594dd] VM Resumed (Lifecycle Event)\n```","commit_id":"1e3143632f30065047ad325d84b93a545aaca4bd"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"65fa5f98860d6d58e46a4b3588edbf1de648bbb9","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":36,"id":"c31d9dee_81b4aca7","updated":"2024-11-20 15:03:32.000000000","message":"still good","commit_id":"e20f152d492abf66a7c5b878a0b35c2360e03bb1"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"fc93f5bf26898f07faaf634cf43bd277aa973240","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":38,"id":"d48765a3_8d0358f1","updated":"2024-11-21 13:00:03.000000000","message":"restating the +2/+W due to the only modification in the commit msg.","commit_id":"9b75726ef35d13f7fa558f4bbe52bd02ee3ad2dd"},{"author":{"_account_id":16207,"name":"ribaudr","display_name":"uggla","email":"rene.ribaud@gmail.com","username":"uggla","status":"Red Hat"},"change_message_id":"fb0fbc872fe3890ef2dce0895e066d000add13a3","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":39,"id":"7d2f4efa_4c883732","updated":"2024-11-25 08:53:38.000000000","message":"recheck","commit_id":"3f3018cde7fc535e5b6a1b38ce591be186ec6dd2"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"a66e01043e2e87884355b8e96b90ea10e128f02e","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":39,"id":"76d9bc8c_43e93527","updated":"2024-11-23 09:16:46.000000000","message":"recheck metadata service timed out in tempest","commit_id":"3f3018cde7fc535e5b6a1b38ce591be186ec6dd2"}],"nova/compute/manager.py":[{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"b5faa1ac37fa9807a6c301b5dfbc46e82efcae30","unresolved":true,"context_lines":[{"line_number":1090,"context_line":"                    \"Share %(share_id)s could not be mounted for \""},{"line_number":1091,"context_line":"                    \"instance %(instance_uuid)s.\","},{"line_number":1092,"context_line":"                    {"},{"line_number":1093,"context_line":"                        \"share_id\": share_info[\"id\"],"},{"line_number":1094,"context_line":"                        \"instance_uuid\": instance.uuid,"},{"line_number":1095,"context_line":"                    },"},{"line_number":1096,"context_line":"                )"}],"source_content_type":"text/x-python","patch_set":14,"id":"ec617146_65a3327d","line":1093,"updated":"2024-01-30 09:48:17.000000000","message":"I see an error here if I simulate a mount failure during compute startup\n\n```\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service Traceback (most recent call last):\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py\", line 810, in run_service\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service     service.start()\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/service.py\", line 162, in start\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service     self.manager.init_host(self.service_ref)\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/compute/manager.py\", line 1684, in init_host\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service     self._init_instance(context, instance)\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/compute/manager.py\", line 1093, in _init_instance\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service     \"share_id\": share_info[\"id\"],\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py\", line 810, in __getitem__\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service     return self.objects[index]\nJan 29 18:29:23 gibi-devstack-aio-jammy nova-compute[163822]: ERROR oslo_service.service TypeError: list indices must be integers or slices, not str\n```\n\nhttps://paste.opendev.org/show/b5lYNqihZgCBECLrN3Td/\n\nYou probably meant `share_mapping[\"id\"]` instead. But this indicate that we are missing a test case.","commit_id":"8910833af2492bf09f016bf4511b9406596100fc"},{"author":{"_account_id":16207,"name":"ribaudr","display_name":"uggla","email":"rene.ribaud@gmail.com","username":"uggla","status":"Red Hat"},"change_message_id":"9a4111ba31209769b9c2239e6169dcfa934bd132","unresolved":false,"context_lines":[{"line_number":1090,"context_line":"                    \"Share %(share_id)s could not be mounted for \""},{"line_number":1091,"context_line":"                    \"instance %(instance_uuid)s.\","},{"line_number":1092,"context_line":"                    {"},{"line_number":1093,"context_line":"                        \"share_id\": share_info[\"id\"],"},{"line_number":1094,"context_line":"                        \"instance_uuid\": instance.uuid,"},{"line_number":1095,"context_line":"                    },"},{"line_number":1096,"context_line":"                )"}],"source_content_type":"text/x-python","patch_set":14,"id":"1def5f3b_f9ae2f08","line":1093,"in_reply_to":"ec617146_65a3327d","updated":"2024-02-16 14:19:06.000000000","message":"I fixed it and I have added a UT checking this.","commit_id":"8910833af2492bf09f016bf4511b9406596100fc"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"b5faa1ac37fa9807a6c301b5dfbc46e82efcae30","unresolved":true,"context_lines":[{"line_number":1085,"context_line":"        for share_mapping in share_info:"},{"line_number":1086,"context_line":"            try:"},{"line_number":1087,"context_line":"                self.mount_share(context, instance, share_mapping)"},{"line_number":1088,"context_line":"            except exception.ShareMountError:"},{"line_number":1089,"context_line":"                LOG.warning("},{"line_number":1090,"context_line":"                    \"Share %(share_id)s could not be mounted for \""},{"line_number":1091,"context_line":"                    \"instance %(instance_uuid)s.\","},{"line_number":1092,"context_line":"                    {"},{"line_number":1093,"context_line":"                        \"share_id\": share_info[\"id\"],"},{"line_number":1094,"context_line":"                        \"instance_uuid\": instance.uuid,"},{"line_number":1095,"context_line":"                    },"},{"line_number":1096,"context_line":"                )"},{"line_number":1097,"context_line":""},{"line_number":1098,"context_line":"        # Instances that are shut down, or in an error state can not be"},{"line_number":1099,"context_line":"        # initialized and are not attempted to be recovered. The exception"}],"source_content_type":"text/x-python","patch_set":14,"id":"3020cf54_c9dc03e5","line":1096,"range":{"start_line":1088,"start_character":0,"end_line":1096,"end_character":17},"updated":"2024-01-30 09:48:17.000000000","message":"After a hypervisor reboot with libvirt virt driver the guest will be in stopped state. Will the compute manager retry the host mount during the start_instance RPC call? or will this error in init_instance cause that the instance is put to ERROR state so start_instance will not be possible?","commit_id":"8910833af2492bf09f016bf4511b9406596100fc"},{"author":{"_account_id":16207,"name":"ribaudr","display_name":"uggla","email":"rene.ribaud@gmail.com","username":"uggla","status":"Red Hat"},"change_message_id":"9a4111ba31209769b9c2239e6169dcfa934bd132","unresolved":true,"context_lines":[{"line_number":1085,"context_line":"        for share_mapping in share_info:"},{"line_number":1086,"context_line":"            try:"},{"line_number":1087,"context_line":"                self.mount_share(context, instance, share_mapping)"},{"line_number":1088,"context_line":"            except exception.ShareMountError:"},{"line_number":1089,"context_line":"                LOG.warning("},{"line_number":1090,"context_line":"                    \"Share %(share_id)s could not be mounted for \""},{"line_number":1091,"context_line":"                    \"instance %(instance_uuid)s.\","},{"line_number":1092,"context_line":"                    {"},{"line_number":1093,"context_line":"                        \"share_id\": share_info[\"id\"],"},{"line_number":1094,"context_line":"                        \"instance_uuid\": instance.uuid,"},{"line_number":1095,"context_line":"                    },"},{"line_number":1096,"context_line":"                )"},{"line_number":1097,"context_line":""},{"line_number":1098,"context_line":"        # Instances that are shut down, or in an error state can not be"},{"line_number":1099,"context_line":"        # initialized and are not attempted to be recovered. The exception"}],"source_content_type":"text/x-python","patch_set":14,"id":"5918cd05_6c5ac60a","line":1096,"range":{"start_line":1088,"start_character":0,"end_line":1096,"end_character":17},"in_reply_to":"3020cf54_c9dc03e5","updated":"2024-02-16 14:19:06.000000000","message":"As of now, if the share is in error, the instances that will use this share will not be started and set to ERROR.","commit_id":"8910833af2492bf09f016bf4511b9406596100fc"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"5fac5bf102d0bb8775afbc366f9523a773baf044","unresolved":true,"context_lines":[{"line_number":1085,"context_line":"        for share_mapping in share_info:"},{"line_number":1086,"context_line":"            try:"},{"line_number":1087,"context_line":"                self.mount_share(context, instance, share_mapping)"},{"line_number":1088,"context_line":"            except exception.ShareMountError:"},{"line_number":1089,"context_line":"                LOG.warning("},{"line_number":1090,"context_line":"                    \"Share %(share_id)s could not be mounted for \""},{"line_number":1091,"context_line":"                    \"instance %(instance_uuid)s.\","},{"line_number":1092,"context_line":"                    {"},{"line_number":1093,"context_line":"                        \"share_id\": share_info[\"id\"],"},{"line_number":1094,"context_line":"                        \"instance_uuid\": instance.uuid,"},{"line_number":1095,"context_line":"                    },"},{"line_number":1096,"context_line":"                )"},{"line_number":1097,"context_line":""},{"line_number":1098,"context_line":"        # Instances that are shut down, or in an error state can not be"},{"line_number":1099,"context_line":"        # initialized and are not attempted to be recovered. The exception"}],"source_content_type":"text/x-python","patch_set":14,"id":"899bc46b_8b6047ed","line":1096,"range":{"start_line":1088,"start_character":0,"end_line":1096,"end_character":17},"in_reply_to":"5918cd05_6c5ac60a","updated":"2024-02-19 16:47:33.000000000","message":"OK, so, if the mount fails here, then the instance will be put into ERROR. That is OK to me.","commit_id":"8910833af2492bf09f016bf4511b9406596100fc"},{"author":{"_account_id":16207,"name":"ribaudr","display_name":"uggla","email":"rene.ribaud@gmail.com","username":"uggla","status":"Red Hat"},"change_message_id":"aff2578b092b2e9dc46f17487aa9a694a2bc2224","unresolved":false,"context_lines":[{"line_number":1085,"context_line":"        for share_mapping in share_info:"},{"line_number":1086,"context_line":"            try:"},{"line_number":1087,"context_line":"                self.mount_share(context, instance, share_mapping)"},{"line_number":1088,"context_line":"            except exception.ShareMountError:"},{"line_number":1089,"context_line":"                LOG.warning("},{"line_number":1090,"context_line":"                    \"Share %(share_id)s could not be mounted for \""},{"line_number":1091,"context_line":"                    \"instance %(instance_uuid)s.\","},{"line_number":1092,"context_line":"                    {"},{"line_number":1093,"context_line":"                        \"share_id\": share_info[\"id\"],"},{"line_number":1094,"context_line":"                        \"instance_uuid\": instance.uuid,"},{"line_number":1095,"context_line":"                    },"},{"line_number":1096,"context_line":"                )"},{"line_number":1097,"context_line":""},{"line_number":1098,"context_line":"        # Instances that are shut down, or in an error state can not be"},{"line_number":1099,"context_line":"        # initialized and are not attempted to be recovered. The exception"}],"source_content_type":"text/x-python","patch_set":14,"id":"1af72508_4fcb3983","line":1096,"range":{"start_line":1088,"start_character":0,"end_line":1096,"end_character":17},"in_reply_to":"899bc46b_8b6047ed","updated":"2024-08-26 14:38:56.000000000","message":"Done","commit_id":"8910833af2492bf09f016bf4511b9406596100fc"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"672ac74da663fc2aadda9f5c21fe8ca87c85b4ce","unresolved":true,"context_lines":[{"line_number":1085,"context_line":"        try:"},{"line_number":1086,"context_line":"            self._mount_all_shares(context, instance, share_info)"},{"line_number":1087,"context_line":"        except exception.ShareMountError as e:"},{"line_number":1088,"context_line":"            LOG.error(e.message)"},{"line_number":1089,"context_line":""},{"line_number":1090,"context_line":"        # Instances that are shut down, or in an error state can not be"},{"line_number":1091,"context_line":"        # initialized and are not attempted to be recovered. The exception"}],"source_content_type":"text/x-python","patch_set":17,"id":"3ce8245f_93e0fb44","line":1088,"updated":"2024-02-28 13:22:36.000000000","message":"as @Sean noted this can go away as you added the mount call to the instance start and reboot codepath","commit_id":"d2f4b276f63429d9e50241e2ec50dcab40bb9942"},{"author":{"_account_id":16207,"name":"ribaudr","display_name":"uggla","email":"rene.ribaud@gmail.com","username":"uggla","status":"Red Hat"},"change_message_id":"aff2578b092b2e9dc46f17487aa9a694a2bc2224","unresolved":false,"context_lines":[{"line_number":1085,"context_line":"        try:"},{"line_number":1086,"context_line":"            self._mount_all_shares(context, instance, share_info)"},{"line_number":1087,"context_line":"        except exception.ShareMountError as e:"},{"line_number":1088,"context_line":"            LOG.error(e.message)"},{"line_number":1089,"context_line":""},{"line_number":1090,"context_line":"        # Instances that are shut down, or in an error state can not be"},{"line_number":1091,"context_line":"        # initialized and are not attempted to be recovered. The exception"}],"source_content_type":"text/x-python","patch_set":17,"id":"447082cf_1923c604","line":1088,"in_reply_to":"2a67dcc9_28f6cc35","updated":"2024-08-26 14:38:56.000000000","message":"Done","commit_id":"d2f4b276f63429d9e50241e2ec50dcab40bb9942"},{"author":{"_account_id":16207,"name":"ribaudr","display_name":"uggla","email":"rene.ribaud@gmail.com","username":"uggla","status":"Red Hat"},"change_message_id":"3f178cb4f561fe6dab750252b7087370732968b3","unresolved":false,"context_lines":[{"line_number":1085,"context_line":"        try:"},{"line_number":1086,"context_line":"            self._mount_all_shares(context, instance, share_info)"},{"line_number":1087,"context_line":"        except exception.ShareMountError as e:"},{"line_number":1088,"context_line":"            LOG.error(e.message)"},{"line_number":1089,"context_line":""},{"line_number":1090,"context_line":"        # Instances that are shut down, or in an error state can not be"},{"line_number":1091,"context_line":"        # initialized and are not attempted to be recovered. The exception"}],"source_content_type":"text/x-python","patch_set":17,"id":"891a652c_70773695","line":1088,"in_reply_to":"3ce8245f_93e0fb44","updated":"2024-03-11 15:44:27.000000000","message":"Done","commit_id":"d2f4b276f63429d9e50241e2ec50dcab40bb9942"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"84edeef9ea6f88d1d000ef2e8ec37919a4246c15","unresolved":true,"context_lines":[{"line_number":1085,"context_line":"        try:"},{"line_number":1086,"context_line":"            self._mount_all_shares(context, instance, share_info)"},{"line_number":1087,"context_line":"        except exception.ShareMountError as e:"},{"line_number":1088,"context_line":"            LOG.error(e.message)"},{"line_number":1089,"context_line":""},{"line_number":1090,"context_line":"        # Instances that are shut down, or in an error state can not be"},{"line_number":1091,"context_line":"        # initialized and are not attempted to be recovered. The exception"}],"source_content_type":"text/x-python","patch_set":17,"id":"2a67dcc9_28f6cc35","line":1088,"in_reply_to":"891a652c_70773695","updated":"2024-05-22 16:28:15.000000000","message":"I guess this was a misunderstanding as Sean thought that the mount moved to the spawn codepath but it is a separate step in the compute manager. As this code calls the driver reboot codepath we missing the mounting now.","commit_id":"d2f4b276f63429d9e50241e2ec50dcab40bb9942"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"84edeef9ea6f88d1d000ef2e8ec37919a4246c15","unresolved":true,"context_lines":[{"line_number":1327,"context_line":"        block_device_info \u003d \\"},{"line_number":1328,"context_line":"            self._get_instance_block_device_info(context, instance)"},{"line_number":1329,"context_line":""},{"line_number":1330,"context_line":"        share_info \u003d self._get_share_info(context, instance)"},{"line_number":1331,"context_line":""},{"line_number":1332,"context_line":"        try:"},{"line_number":1333,"context_line":"            self.driver.resume_state_on_host_boot("},{"line_number":1334,"context_line":"                context, instance, net_info, share_info, block_device_info)"},{"line_number":1335,"context_line":"        except NotImplementedError:"},{"line_number":1336,"context_line":"            LOG.warning(\u0027Hypervisor driver does not support \u0027"},{"line_number":1337,"context_line":"                        \u0027resume guests\u0027, instance\u003dinstance)"}],"source_content_type":"text/x-python","patch_set":22,"id":"a45b2e8f_a58ed588","line":1334,"range":{"start_line":1330,"start_character":0,"end_line":1334,"end_character":75},"updated":"2024-05-22 16:28:15.000000000","message":"At this point mounts might missing due to the hypervisor reboot, but as far as I see the code does not try to mount the shares before the reboot. So the instance will loose the share. \n\nHave you tested resume_guests_state_on_host_boot with manila shares?","commit_id":"54fbd00ef0c68652554e799e1d4f78d25bf75309"},{"author":{"_account_id":16207,"name":"ribaudr","display_name":"uggla","email":"rene.ribaud@gmail.com","username":"uggla","status":"Red Hat"},"change_message_id":"e00aeebcabe2ff89f1affdd3d23c1f01a76c5300","unresolved":false,"context_lines":[{"line_number":1327,"context_line":"        block_device_info \u003d \\"},{"line_number":1328,"context_line":"            self._get_instance_block_device_info(context, instance)"},{"line_number":1329,"context_line":""},{"line_number":1330,"context_line":"        share_info \u003d self._get_share_info(context, instance)"},{"line_number":1331,"context_line":""},{"line_number":1332,"context_line":"        try:"},{"line_number":1333,"context_line":"            self.driver.resume_state_on_host_boot("},{"line_number":1334,"context_line":"                context, instance, net_info, share_info, block_device_info)"},{"line_number":1335,"context_line":"        except NotImplementedError:"},{"line_number":1336,"context_line":"            LOG.warning(\u0027Hypervisor driver does not support \u0027"},{"line_number":1337,"context_line":"                        \u0027resume guests\u0027, instance\u003dinstance)"}],"source_content_type":"text/x-python","patch_set":22,"id":"346935ef_7125b08d","line":1334,"range":{"start_line":1330,"start_character":0,"end_line":1334,"end_character":75},"in_reply_to":"205d1e51_483a0f16","updated":"2024-07-16 12:46:02.000000000","message":"Done","commit_id":"54fbd00ef0c68652554e799e1d4f78d25bf75309"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"3f18047a45191b89b4737822feeadbffe95c3a89","unresolved":true,"context_lines":[{"line_number":1327,"context_line":"        block_device_info \u003d \\"},{"line_number":1328,"context_line":"            self._get_instance_block_device_info(context, instance)"},{"line_number":1329,"context_line":""},{"line_number":1330,"context_line":"        share_info \u003d self._get_share_info(context, instance)"},{"line_number":1331,"context_line":""},{"line_number":1332,"context_line":"        try:"},{"line_number":1333,"context_line":"            self.driver.resume_state_on_host_boot("},{"line_number":1334,"context_line":"                context, instance, net_info, share_info, block_device_info)"},{"line_number":1335,"context_line":"        except NotImplementedError:"},{"line_number":1336,"context_line":"            LOG.warning(\u0027Hypervisor driver does not support \u0027"},{"line_number":1337,"context_line":"                        \u0027resume guests\u0027, instance\u003dinstance)"}],"source_content_type":"text/x-python","patch_set":22,"id":"205d1e51_483a0f16","line":1334,"range":{"start_line":1330,"start_character":0,"end_line":1334,"end_character":75},"in_reply_to":"6defb489_2ba78e87","updated":"2024-07-03 15:50:25.000000000","message":"René, the problem is with a host reboot : in that case, the mount will disappear after the host reboot, and we need to ensure that the mounts are present.","commit_id":"54fbd00ef0c68652554e799e1d4f78d25bf75309"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"cfeaca47aa99bea06bb02734286084540f72bb8b","unresolved":true,"context_lines":[{"line_number":1327,"context_line":"        block_device_info \u003d \\"},{"line_number":1328,"context_line":"            self._get_instance_block_device_info(context, instance)"},{"line_number":1329,"context_line":""},{"line_number":1330,"context_line":"        share_info \u003d self._get_share_info(context, instance)"},{"line_number":1331,"context_line":""},{"line_number":1332,"context_line":"        try:"},{"line_number":1333,"context_line":"            self.driver.resume_state_on_host_boot("},{"line_number":1334,"context_line":"                context, instance, net_info, share_info, block_device_info)"},{"line_number":1335,"context_line":"        except NotImplementedError:"},{"line_number":1336,"context_line":"            LOG.warning(\u0027Hypervisor driver does not support \u0027"},{"line_number":1337,"context_line":"                        \u0027resume guests\u0027, instance\u003dinstance)"}],"source_content_type":"text/x-python","patch_set":22,"id":"6216113c_5cd0feb8","line":1334,"range":{"start_line":1330,"start_character":0,"end_line":1334,"end_character":75},"in_reply_to":"a45b2e8f_a58ed588","updated":"2024-05-28 13:13:00.000000000","message":"Lets check if spawn will do the mount or only the compute manager","commit_id":"54fbd00ef0c68652554e799e1d4f78d25bf75309"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"cfeaca47aa99bea06bb02734286084540f72bb8b","unresolved":true,"context_lines":[{"line_number":1327,"context_line":"        block_device_info \u003d \\"},{"line_number":1328,"context_line":"            self._get_instance_block_device_info(context, instance)"},{"line_number":1329,"context_line":""},{"line_number":1330,"context_line":"        share_info \u003d self._get_share_info(context, instance)"},{"line_number":1331,"context_line":""},{"line_number":1332,"context_line":"        try:"},{"line_number":1333,"context_line":"            self.driver.resume_state_on_host_boot("},{"line_number":1334,"context_line":"                context, instance, net_info, share_info, block_device_info)"},{"line_number":1335,"context_line":"        except NotImplementedError:"},{"line_number":1336,"context_line":"            LOG.warning(\u0027Hypervisor driver does not support \u0027"},{"line_number":1337,"context_line":"                        \u0027resume guests\u0027, instance\u003dinstance)"}],"source_content_type":"text/x-python","patch_set":22,"id":"ed55bb0f_070aa6b5","line":1334,"range":{"start_line":1330,"start_character":0,"end_line":1334,"end_character":75},"in_reply_to":"a45b2e8f_a58ed588","updated":"2024-05-28 13:13:00.000000000","message":"lets make sure that we have a mount share call in this code path. If that is in the driver call that is fine, but there are other lifecycle operations where the mount call is in the compute manager.","commit_id":"54fbd00ef0c68652554e799e1d4f78d25bf75309"},{"author":{"_account_id":16207,"name":"ribaudr","display_name":"uggla","email":"rene.ribaud@gmail.com","username":"uggla","status":"Red Hat"},"change_message_id":"4a1a90b542c662d7d380938ac5690467a3168d82","unresolved":true,"context_lines":[{"line_number":1327,"context_line":"        block_device_info \u003d \\"},{"line_number":1328,"context_line":"            self._get_instance_block_device_info(context, instance)"},{"line_number":1329,"context_line":""},{"line_number":1330,"context_line":"        share_info \u003d self._get_share_info(context, instance)"},{"line_number":1331,"context_line":""},{"line_number":1332,"context_line":"        try:"},{"line_number":1333,"context_line":"            self.driver.resume_state_on_host_boot("},{"line_number":1334,"context_line":"                context, instance, net_info, share_info, block_device_info)"},{"line_number":1335,"context_line":"        except NotImplementedError:"},{"line_number":1336,"context_line":"            LOG.warning(\u0027Hypervisor driver does not support \u0027"},{"line_number":1337,"context_line":"                        \u0027resume guests\u0027, instance\u003dinstance)"}],"source_content_type":"text/x-python","patch_set":22,"id":"6defb489_2ba78e87","line":1334,"range":{"start_line":1330,"start_character":0,"end_line":1334,"end_character":75},"in_reply_to":"ed55bb0f_070aa6b5","updated":"2024-06-03 14:58:17.000000000","message":"I have reviewed all the cases, I suggest we do it together to make sure I did not miss something. But it appears to me that it is correct.","commit_id":"54fbd00ef0c68652554e799e1d4f78d25bf75309"},{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"9c1e3b31d22241b1e45f337f15cac3604f0a8d24","unresolved":true,"context_lines":[{"line_number":1328,"context_line":"            self._get_instance_block_device_info(context, instance)"},{"line_number":1329,"context_line":""},{"line_number":1330,"context_line":"        share_info \u003d self._get_share_info(context, instance)"},{"line_number":1331,"context_line":"        self._mount_all_shares(context, instance, share_info)"},{"line_number":1332,"context_line":""},{"line_number":1333,"context_line":"        try:"},{"line_number":1334,"context_line":"            self.driver.resume_state_on_host_boot("}],"source_content_type":"text/x-python","patch_set":34,"id":"7dfd8c8e_4680959d","line":1331,"updated":"2024-11-13 14:38:06.000000000","message":"fwiw, I tested and we get an exception as the context doesn\u0027t have an admin context  : \n\n```\n\n\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service [None req-6171d2da-ec69-4872-8b01-0d3d37bad70e None None] Error starting thread.: TypeError: Token.__init__() missing 1 required positional argument: \u0027token\u0027\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service Traceback (most recent call last):\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py\", line 810, in run_service\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     service.start()\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/service.py\", line 166, in start\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     self.manager.init_host(self.service_ref)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/compute/manager.py\", line 1677, in init_host\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     self._init_instance(context, instance)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/compute/manager.py\", line 1322, in _init_instance\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     self._resume_guests_state(context, instance, net_info)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/compute/manager.py\", line 1331, in _resume_guests_state\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     self._mount_all_shares(context, instance, share_info)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/exception_wrapper.py\", line 65, in wrapped\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     with excutils.save_and_reraise_exception():\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py\", line 227, in __exit__\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     self.force_reraise()\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py\", line 200, in force_reraise\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     raise self.value\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/exception_wrapper.py\", line 63, in wrapped\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     return f(self, context, *args, **kw)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/compute/manager.py\", line 4913, in _mount_all_shares\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     self._mount_share(context, instance, share_mapping)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/exception_wrapper.py\", line 65, in wrapped\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     with excutils.save_and_reraise_exception():\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py\", line 227, in __exit__\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     self.force_reraise()\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py\", line 200, in force_reraise\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     raise self.value\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/exception_wrapper.py\", line 63, in wrapped\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     return f(self, context, *args, **kw)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/compute/manager.py\", line 4953, in _mount_share\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     _mount_share(context, instance, share_mapping)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py\", line 412, in inner\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     return f(*args, **kwargs)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/compute/manager.py\", line 4930, in _mount_share\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     share_mapping.enhance_with_ceph_credentials(context)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/objects/share_mapping.py\", line 155, in enhance_with_ceph_credentials\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     access \u003d manila_api.API().get_access(\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/share/manila.py\", line 128, in wrapper\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     res \u003d method(self, *args, **kwargs)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/share/manila.py\", line 146, in wrapper\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     res \u003d method(self, *args, **kwargs)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/share/manila.py\", line 244, in get_access\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     access_list \u003d _manilaclient(\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/share/manila.py\", line 42, in _manilaclient\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     return utils.get_sdk_adapter(\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/nova/nova/utils.py\", line 1007, in get_sdk_adapter\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     conn \u003d connection.Connection(\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/openstack/connection.py\", line 449, in __init__\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     super().__init__(\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/openstack/cloud/_image.py\", line 21, in __init__\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     super().__init__(*args, **kwargs)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/openstack/cloud/_network_common.py\", line 32, in __init__\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     super().__init__(*args, **kwargs)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/openstack/cloud/openstackcloud.py\", line 192, in __init__\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     self.config \u003d openstack.config.get_cloud_region(\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/openstack/config/__init__.py\", line 41, in get_cloud_region\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     return config.get_one(options\u003dparsed_options, **kwargs)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/openstack/config/loader.py\", line 1294, in get_one\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     auth_plugin \u003d loader.load_from_options(**config[\u0027auth\u0027])\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/keystoneauth1/loading/identity.py\", line 103, in load_from_options\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     return super(BaseV3Loader, self).load_from_options(**kwargs)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/keystoneauth1/loading/base.py\", line 162, in load_from_options\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     return self.create_plugin(**kwargs)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service   File \"/opt/stack/data/venv/lib/python3.10/site-packages/keystoneauth1/loading/base.py\", line 123, in create_plugin\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service     return self.plugin_class(**kwargs)\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service TypeError: Token.__init__() missing 1 required positional argument: \u0027token\u0027\n    Nov 13 09:30:13 sbauza-manila2 nova-compute[828]: ERROR oslo_service.service \n\n\n```\n\nWe need to fix that or throw that patch later in the series.","commit_id":"ece68a4df2d7ca206cbad74dc48405a3c520c1c7"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"b1f3dc1050347786ae1946bdca620f0232d36330","unresolved":true,"context_lines":[{"line_number":1328,"context_line":"            self._get_instance_block_device_info(context, instance)"},{"line_number":1329,"context_line":""},{"line_number":1330,"context_line":"        share_info \u003d self._get_share_info(context, instance)"},{"line_number":1331,"context_line":"        self._mount_all_shares(context, instance, share_info)"},{"line_number":1332,"context_line":""},{"line_number":1333,"context_line":"        try:"},{"line_number":1334,"context_line":"            self.driver.resume_state_on_host_boot("}],"source_content_type":"text/x-python","patch_set":34,"id":"35241719_425939d5","line":1331,"in_reply_to":"03a487cd_75e9a12b","updated":"2024-11-14 10:12:33.000000000","message":"@smooney@redhat.com this is the case of resume guest state on hypevisor reboot. So we don\u0027t have the user token available to nova-compute, as there was no user request associated to the trigger, the startup of the nova-compute service. \n\nIn the nova\u0027s manila client code we have to modes to talk to manila, either via the user token, or via nova\u0027s service token. \n\nThe share mount logic needs to read the CEPH credentials from manila to be able to mount the share during nova-compute startup (if resume guest state is configured). So in this specific scenario nova needs to use the service token. But in a normal VM hard reboot scenario the user token can be used.","commit_id":"ece68a4df2d7ca206cbad74dc48405a3c520c1c7"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"a9af8eae41af738f47daae91597462cbb6ba842f","unresolved":true,"context_lines":[{"line_number":1328,"context_line":"            self._get_instance_block_device_info(context, instance)"},{"line_number":1329,"context_line":""},{"line_number":1330,"context_line":"        share_info \u003d self._get_share_info(context, instance)"},{"line_number":1331,"context_line":"        self._mount_all_shares(context, instance, share_info)"},{"line_number":1332,"context_line":""},{"line_number":1333,"context_line":"        try:"},{"line_number":1334,"context_line":"            self.driver.resume_state_on_host_boot("}],"source_content_type":"text/x-python","patch_set":34,"id":"9afbaf34_395a9d99","line":1331,"in_reply_to":"35241719_425939d5","updated":"2024-11-15 10:24:33.000000000","message":"@smooney@redhat.com I also summarized my views in https://review.opendev.org/c/openstack/nova/+/925277/11#message-0b59e302be6bffdc887c186b5fe67a89ba8ef3f3 Could you check that commit please? We would need a 3rd core there as I wrote most of the logic there.","commit_id":"ece68a4df2d7ca206cbad74dc48405a3c520c1c7"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"75f7bf22388c0212a465de2505c6556d77577e8f","unresolved":true,"context_lines":[{"line_number":1328,"context_line":"            self._get_instance_block_device_info(context, instance)"},{"line_number":1329,"context_line":""},{"line_number":1330,"context_line":"        share_info \u003d self._get_share_info(context, instance)"},{"line_number":1331,"context_line":"        self._mount_all_shares(context, instance, share_info)"},{"line_number":1332,"context_line":""},{"line_number":1333,"context_line":"        try:"},{"line_number":1334,"context_line":"            self.driver.resume_state_on_host_boot("}],"source_content_type":"text/x-python","patch_set":34,"id":"03a487cd_75e9a12b","line":1331,"in_reply_to":"7dfd8c8e_4680959d","updated":"2024-11-13 15:22:22.000000000","message":"we are not ment to be using admin credetials to talke to manially we are ment to be useing the user token + a service toke to talk to manilla as far as i am aware.\n\nso that confuses me, am i wrong about that?","commit_id":"ece68a4df2d7ca206cbad74dc48405a3c520c1c7"}],"nova/virt/libvirt/driver.py":[{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"3f18047a45191b89b4737822feeadbffe95c3a89","unresolved":true,"context_lines":[{"line_number":4402,"context_line":"        # Be as absolute as possible about getting it back into"},{"line_number":4403,"context_line":"        # a known and running state."},{"line_number":4404,"context_line":"        self._hard_reboot(context, instance, network_info,"},{"line_number":4405,"context_line":"            share_info, block_device_info"},{"line_number":4406,"context_line":"        )"},{"line_number":4407,"context_line":""},{"line_number":4408,"context_line":"    def rescue(self, context, instance, network_info, image_meta,"}],"source_content_type":"text/x-python","patch_set":26,"id":"3519aba2_28ac8404","line":4405,"updated":"2024-07-03 15:50:25.000000000","message":"here, see, you\u0027re passing a share info without checking the mount state, compared to the poweron case.","commit_id":"f0be8d95f6d42fe06db5d21f951209618ffc3fb0"},{"author":{"_account_id":16207,"name":"ribaudr","display_name":"uggla","email":"rene.ribaud@gmail.com","username":"uggla","status":"Red Hat"},"change_message_id":"e00aeebcabe2ff89f1affdd3d23c1f01a76c5300","unresolved":false,"context_lines":[{"line_number":4402,"context_line":"        # Be as absolute as possible about getting it back into"},{"line_number":4403,"context_line":"        # a known and running state."},{"line_number":4404,"context_line":"        self._hard_reboot(context, instance, network_info,"},{"line_number":4405,"context_line":"            share_info, block_device_info"},{"line_number":4406,"context_line":"        )"},{"line_number":4407,"context_line":""},{"line_number":4408,"context_line":"    def rescue(self, context, instance, network_info, image_meta,"}],"source_content_type":"text/x-python","patch_set":26,"id":"98d62d9d_4443e1f5","line":4405,"in_reply_to":"3519aba2_28ac8404","updated":"2024-07-16 12:46:02.000000000","message":"Acknowledged","commit_id":"f0be8d95f6d42fe06db5d21f951209618ffc3fb0"}]}
