)]}'
{"/COMMIT_MSG":[{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"4489fe9b6db69d32e9c09e62eb3a5a6fca569a67","unresolved":true,"context_lines":[{"line_number":4,"context_line":"Commit:     digvijay2016 \u003cdigvijay.ukirde@in.ibm.com\u003e"},{"line_number":5,"context_line":"CommitDate: 2021-10-23 15:12:52 -0400"},{"line_number":6,"context_line":""},{"line_number":7,"context_line":"Enabling Copy-on-write mode in GPFS NFS driver"},{"line_number":8,"context_line":""},{"line_number":9,"context_line":"IBM Spectrum Scale cinder driver (GPFS) support copy-on-write feature"},{"line_number":10,"context_line":"in all the configuration. Resolving the bugs mentioned below will"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":4,"id":"253a3d3c_f8dbe557","line":7,"range":{"start_line":7,"start_character":0,"end_line":7,"end_character":8},"updated":"2021-10-25 17:15:57.000000000","message":"nit: \"Enable\" or \"Enabled\" (not \"enabling\")\n\nBut wouldn\u0027t it be more accurate to state you fixed this? (\"Fixed copy-on-write mode in GPFS NFS driver\")","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"26ef7911ee1532c1b92ce8c6c475f899d29feea7","unresolved":false,"context_lines":[{"line_number":4,"context_line":"Commit:     digvijay2016 \u003cdigvijay.ukirde@in.ibm.com\u003e"},{"line_number":5,"context_line":"CommitDate: 2021-10-23 15:12:52 -0400"},{"line_number":6,"context_line":""},{"line_number":7,"context_line":"Enabling Copy-on-write mode in GPFS NFS driver"},{"line_number":8,"context_line":""},{"line_number":9,"context_line":"IBM Spectrum Scale cinder driver (GPFS) support copy-on-write feature"},{"line_number":10,"context_line":"in all the configuration. Resolving the bugs mentioned below will"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":4,"id":"0f0e050c_c150dee0","line":7,"range":{"start_line":7,"start_character":0,"end_line":7,"end_character":8},"in_reply_to":"253a3d3c_f8dbe557","updated":"2021-11-02 15:43:22.000000000","message":"Addressed in the latest patchset. Done","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"}],"/PATCHSET_LEVEL":[{"author":{"_account_id":32036,"name":"katari manoj kumar","email":"katkumar@in.ibm.com","username":"katarimanojkumar"},"change_message_id":"33bd8c0ea0ae690f101a26d11cd838a6276b4b27","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"d67252a7_0cf11175","updated":"2021-10-21 10:58:47.000000000","message":"Pls check and add release notes.","commit_id":"6b04d740267eb18a98ca8f37c6cb7949d2b9a342"},{"author":{"_account_id":19904,"name":"Digvijay Ukirde","email":"digvijay.ukirde@in.ibm.com","username":"digvijay2016"},"change_message_id":"48a8504f872ac339777c842f835edb383eba455c","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"9b035c7c_e0db3543","updated":"2021-10-19 07:43:57.000000000","message":"recheck","commit_id":"6b04d740267eb18a98ca8f37c6cb7949d2b9a342"},{"author":{"_account_id":19904,"name":"Digvijay Ukirde","email":"digvijay.ukirde@in.ibm.com","username":"digvijay2016"},"change_message_id":"9b8b92ce3a889d6aac942375a3b7364dfe66a609","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"c05ce876_a71c733b","in_reply_to":"d67252a7_0cf11175","updated":"2021-10-24 13:40:34.000000000","message":"Thanks for the review. Added release notes.","commit_id":"6b04d740267eb18a98ca8f37c6cb7949d2b9a342"},{"author":{"_account_id":27615,"name":"Rajat Dhasmana","email":"rajatdhasmana@gmail.com","username":"whoami-rajat"},"change_message_id":"628b37d4c872df8b3fc783d186d1ded0c5a371d6","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":4,"id":"b144e3ab_9306f32e","updated":"2021-10-28 14:33:42.000000000","message":"Thanks Alan for clarifying","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":32036,"name":"katari manoj kumar","email":"katkumar@in.ibm.com","username":"katarimanojkumar"},"change_message_id":"1844481a4a82b4c6bd7775f451d6e1ab3c82f7e6","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":4,"id":"b3ad76c5_5dca049d","updated":"2021-10-25 04:59:12.000000000","message":"Thanks for addressing my comments. LGTM ","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"cd30efad9a00b744fd9cb1689b5d6ff07da60926","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":7,"id":"6813a676_d2511eba","updated":"2021-11-01 13:52:34.000000000","message":"The code and unit test changes lgtm, but I have a couple of concerns with the release note, especially the confusion between the LP bug description and the statement that this patch fixes a delete_volume() failure when the provider_location is empty.\n\nI also like the release note file name suggested by Rajat, or something like \"gpfs-fix-nfs-cow\". The idea is to mention the driver in the release note\u0027s file name.","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"26ef7911ee1532c1b92ce8c6c475f899d29feea7","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":7,"id":"d146d099_2a3fa7f6","in_reply_to":"6813a676_d2511eba","updated":"2021-11-02 15:43:22.000000000","message":"Thanks for the review Alan. \n1. I will rename the release notes file name as \"bug-gpfs-fix-nfs-cow.yaml\". Is that fine?\n2. Replacing release notes content as given below. Is that okay?\n\n---\nfixes:\n  - |\n    IBM GPFS driver `Bug #1947134 \u003chttps://bugs.launchpad.net/cinder/+bug/1947134\u003e`_: Fixed\n    the initialization of GPFS NFS driver when gpfs_images_share_mode\n    is set to copy_on_write by correcting _same_filesystem functionality.\n  - |\n    IBM GPFS driver `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed\n    the volume creation issue in GPFS NFS driver when gpfs_images_share_mode\n    is set to copy_on_write.\n    the volume deletion issue when volume provider location provides null\n    for volume in error state.","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"4bc2ad4f11ce3f44011b5165142a26be340cdcfe","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":7,"id":"f4a36799_179d8f38","in_reply_to":"d146d099_2a3fa7f6","updated":"2021-11-02 23:12:43.000000000","message":"Addressed in the latest patchset. Done","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"3e14f3fc164a6d01d93e86dea0e9abfa0ae33958","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":8,"id":"b0066501_d89915fe","updated":"2021-11-09 17:05:19.000000000","message":"Because this patch fixes a couple of bugs, you might consider proposing a backport to the stable branches, starting with stable/xena","commit_id":"dcc191646985fa0be76ee763935ef37a8456d1e6"},{"author":{"_account_id":5314,"name":"Brian Rosmaita","email":"rosmaita.fossdev@gmail.com","username":"brian-rosmaita"},"change_message_id":"82755b8da6c78b5d0ca01fa596bb3fcba4bf7379","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":8,"id":"3be54a4d_d6d80f76","updated":"2021-11-08 19:56:48.000000000","message":"Code and tests look good, and the CI is passing.","commit_id":"dcc191646985fa0be76ee763935ef37a8456d1e6"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"2dd21e111055b877723c71f69c328147d739a7d2","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":8,"id":"0371cb0c_07b99803","updated":"2021-11-05 17:36:56.000000000","message":"IBM GPFS CI still needs to pass.","commit_id":"dcc191646985fa0be76ee763935ef37a8456d1e6"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"da05f1aea1f2f8337e0b2efa4190e1515e855dcf","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":8,"id":"a1345184_8ebefe19","updated":"2021-11-03 13:43:22.000000000","message":"lgtm, but \"IBM GPFS CI\" needs to pass.","commit_id":"dcc191646985fa0be76ee763935ef37a8456d1e6"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"d3a9dcef546e070810c215f7f44a3399fd8b3825","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":8,"id":"9d9495b4_23719d47","updated":"2021-11-03 05:03:11.000000000","message":"recheck","commit_id":"dcc191646985fa0be76ee763935ef37a8456d1e6"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"2f6b22f42c27869b2f85c8d7c8f2555cdb9aec95","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":8,"id":"162dd2a1_d1859b2e","in_reply_to":"0371cb0c_07b99803","updated":"2021-11-05 17:46:37.000000000","message":"Yes, There are few issues in internal CI system. I am working on it to fix it. It will be resolved by this week. In the meantime we want to reach out to Reviewer Rajat in addressing his review comment. If you have irc id of reviewer Rajat, please let us know. Thanks","commit_id":"dcc191646985fa0be76ee763935ef37a8456d1e6"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"8061b40afc0081b71a2806d4c2b9b75603b06df3","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":8,"id":"20c32e8c_473a568f","in_reply_to":"162dd2a1_d1859b2e","updated":"2021-11-08 02:53:41.000000000","message":"IBM GPFS CI is successful now","commit_id":"dcc191646985fa0be76ee763935ef37a8456d1e6"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"5c7f2d242a028a12e586aa9cf4b28b2d3c77a867","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":8,"id":"0df809d1_2561b20a","in_reply_to":"a1345184_8ebefe19","updated":"2021-11-04 14:04:09.000000000","message":"Thanks Alan. Can you share Rajat irc and email id to contact and resolve remaining comments?","commit_id":"dcc191646985fa0be76ee763935ef37a8456d1e6"}],"cinder/tests/unit/volume/drivers/test_gpfs.py":[{"author":{"_account_id":32036,"name":"katari manoj kumar","email":"katkumar@in.ibm.com","username":"katarimanojkumar"},"change_message_id":"33bd8c0ea0ae690f101a26d11cd838a6276b4b27","unresolved":true,"context_lines":[{"line_number":564,"context_line":"        self.override_config(\u0027gpfs_images_share_mode\u0027, \u0027copy_on_write\u0027,"},{"line_number":565,"context_line":"                             conf.SHARED_CONF_GROUP)"},{"line_number":566,"context_line":"        with mock.patch(\u0027cinder.volume.drivers.ibm.gpfs.GPFSDriver.\u0027"},{"line_number":567,"context_line":"                        \u0027_same_filesystem\u0027, return_value\u003dFalse):"},{"line_number":568,"context_line":"            self.assertRaises(exception.VolumeBackendAPIException,"},{"line_number":569,"context_line":"                              self.driver.check_for_setup_error)"},{"line_number":570,"context_line":""}],"source_content_type":"text/x-python","patch_set":3,"id":"97a4d43d_8734fcc8","line":567,"range":{"start_line":567,"start_character":0,"end_line":567,"end_character":25},"updated":"2021-10-21 10:58:47.000000000","message":"i see no UT for method _same_filesystem()","commit_id":"6b04d740267eb18a98ca8f37c6cb7949d2b9a342"},{"author":{"_account_id":19904,"name":"Digvijay Ukirde","email":"digvijay.ukirde@in.ibm.com","username":"digvijay2016"},"change_message_id":"9b8b92ce3a889d6aac942375a3b7364dfe66a609","unresolved":false,"context_lines":[{"line_number":564,"context_line":"        self.override_config(\u0027gpfs_images_share_mode\u0027, \u0027copy_on_write\u0027,"},{"line_number":565,"context_line":"                             conf.SHARED_CONF_GROUP)"},{"line_number":566,"context_line":"        with mock.patch(\u0027cinder.volume.drivers.ibm.gpfs.GPFSDriver.\u0027"},{"line_number":567,"context_line":"                        \u0027_same_filesystem\u0027, return_value\u003dFalse):"},{"line_number":568,"context_line":"            self.assertRaises(exception.VolumeBackendAPIException,"},{"line_number":569,"context_line":"                              self.driver.check_for_setup_error)"},{"line_number":570,"context_line":""}],"source_content_type":"text/x-python","patch_set":3,"id":"05b08cb2_f2d1b0fb","line":567,"range":{"start_line":567,"start_character":0,"end_line":567,"end_character":25},"in_reply_to":"97a4d43d_8734fcc8","updated":"2021-10-24 13:40:34.000000000","message":"Done","commit_id":"6b04d740267eb18a98ca8f37c6cb7949d2b9a342"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"4489fe9b6db69d32e9c09e62eb3a5a6fca569a67","unresolved":true,"context_lines":[{"line_number":154,"context_line":"                          self.driver._check_gpfs_state)"},{"line_number":155,"context_line":""},{"line_number":156,"context_line":"    def test_same_filesystem_ok(self):"},{"line_number":157,"context_line":"        self.assertTrue(self.driver._same_filesystem(\u0027/home\u0027, \u0027/root\u0027))"},{"line_number":158,"context_line":""},{"line_number":159,"context_line":"    @mock.patch(\u0027cinder.utils.execute\u0027)"},{"line_number":160,"context_line":"    def test_same_filesystem_failed(self, mock_exec):"}],"source_content_type":"text/x-python","patch_set":4,"id":"f97d798e_985c3681","line":157,"updated":"2021-10-25 17:15:57.000000000","message":"What is the purpose of this test? Under what circumstances are you sure that /home and /root are in fact on the same file system? It seems like you are making an assumption about the file systems of the machine that is running the unit tests.","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"26ef7911ee1532c1b92ce8c6c475f899d29feea7","unresolved":false,"context_lines":[{"line_number":154,"context_line":"                          self.driver._check_gpfs_state)"},{"line_number":155,"context_line":""},{"line_number":156,"context_line":"    def test_same_filesystem_ok(self):"},{"line_number":157,"context_line":"        self.assertTrue(self.driver._same_filesystem(\u0027/home\u0027, \u0027/root\u0027))"},{"line_number":158,"context_line":""},{"line_number":159,"context_line":"    @mock.patch(\u0027cinder.utils.execute\u0027)"},{"line_number":160,"context_line":"    def test_same_filesystem_failed(self, mock_exec):"}],"source_content_type":"text/x-python","patch_set":4,"id":"a6871f8a_87ad6bff","line":157,"in_reply_to":"021f723e_7de467dd","updated":"2021-11-02 15:43:22.000000000","message":"Addressed in the latest patchset. Done","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"5001e08fb1b7aa50b86058db5c23d34780659be4","unresolved":true,"context_lines":[{"line_number":154,"context_line":"                          self.driver._check_gpfs_state)"},{"line_number":155,"context_line":""},{"line_number":156,"context_line":"    def test_same_filesystem_ok(self):"},{"line_number":157,"context_line":"        self.assertTrue(self.driver._same_filesystem(\u0027/home\u0027, \u0027/root\u0027))"},{"line_number":158,"context_line":""},{"line_number":159,"context_line":"    @mock.patch(\u0027cinder.utils.execute\u0027)"},{"line_number":160,"context_line":"    def test_same_filesystem_failed(self, mock_exec):"}],"source_content_type":"text/x-python","patch_set":4,"id":"021f723e_7de467dd","line":157,"in_reply_to":"a1750b35_3d937300","updated":"2021-10-26 16:30:29.000000000","message":"Well, you could mock gpfs_execute and set the mock\u0027s return_value to simulate the (err, out) tuple. And set the \u0027out\u0027 value to return the two path inputs, maybe like this:\n\n  return (0, \"%s\\n%s\\n\" % (path1, path2))\n\nSo it basically returns the exact same two paths, but in a form the _same_filesystem() function will parse in order to return the boolean result. For a bonus, you could use DDT to have it test both conditions (same filesystem is True, and False test cases).","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":19904,"name":"Digvijay Ukirde","email":"digvijay.ukirde@in.ibm.com","username":"digvijay2016"},"change_message_id":"ae855c0a5ac70914947dcc738b535eb441bcee4b","unresolved":true,"context_lines":[{"line_number":154,"context_line":"                          self.driver._check_gpfs_state)"},{"line_number":155,"context_line":""},{"line_number":156,"context_line":"    def test_same_filesystem_ok(self):"},{"line_number":157,"context_line":"        self.assertTrue(self.driver._same_filesystem(\u0027/home\u0027, \u0027/root\u0027))"},{"line_number":158,"context_line":""},{"line_number":159,"context_line":"    @mock.patch(\u0027cinder.utils.execute\u0027)"},{"line_number":160,"context_line":"    def test_same_filesystem_failed(self, mock_exec):"}],"source_content_type":"text/x-python","patch_set":4,"id":"a1750b35_3d937300","line":157,"in_reply_to":"f97d798e_985c3681","updated":"2021-10-26 16:05:55.000000000","message":"Thanks Alan for the review. I didn\u0027t find any way to make sure both the paths are on the same filesystem. The fake paths will result into test case failure so I assumption according to the CI setup. Can you suggest a way.","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"4489fe9b6db69d32e9c09e62eb3a5a6fca569a67","unresolved":true,"context_lines":[{"line_number":1269,"context_line":"                                                  mock_set_rw_permission,"},{"line_number":1270,"context_line":"                                                  mock_resize_volume_file):"},{"line_number":1271,"context_line":"        mock_is_cloneable.return_value \u003d (True, \u0027test\u0027, self.images_dir)"},{"line_number":1272,"context_line":"        mock_local_path.return_value \u003d self.volumes_path"},{"line_number":1273,"context_line":"        mock_is_gpfs_parent_file.return_value \u003d False"},{"line_number":1274,"context_line":"        mock_qemu_img_info.return_value \u003d self._fake_qemu_raw_image_info(\u0027\u0027)"},{"line_number":1275,"context_line":"        volume \u003d self._fake_volume()"}],"source_content_type":"text/x-python","patch_set":4,"id":"fc67257a_aa445621","line":1272,"updated":"2021-10-25 17:15:57.000000000","message":"But in the new code, for raw volumes it should be using self._get_volume_path() and not self.local_path().\n\nThis needs to be updated to test the new code. You may want to verify _get_volume_path() is called, and local_path is *not* called.","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"26ef7911ee1532c1b92ce8c6c475f899d29feea7","unresolved":false,"context_lines":[{"line_number":1269,"context_line":"                                                  mock_set_rw_permission,"},{"line_number":1270,"context_line":"                                                  mock_resize_volume_file):"},{"line_number":1271,"context_line":"        mock_is_cloneable.return_value \u003d (True, \u0027test\u0027, self.images_dir)"},{"line_number":1272,"context_line":"        mock_local_path.return_value \u003d self.volumes_path"},{"line_number":1273,"context_line":"        mock_is_gpfs_parent_file.return_value \u003d False"},{"line_number":1274,"context_line":"        mock_qemu_img_info.return_value \u003d self._fake_qemu_raw_image_info(\u0027\u0027)"},{"line_number":1275,"context_line":"        volume \u003d self._fake_volume()"}],"source_content_type":"text/x-python","patch_set":4,"id":"29e36aea_568499e0","line":1272,"in_reply_to":"fc67257a_aa445621","updated":"2021-11-02 15:43:22.000000000","message":"Addressed in the latest patchset. Done","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"}],"cinder/volume/drivers/ibm/gpfs.py":[{"author":{"_account_id":27615,"name":"Rajat Dhasmana","email":"rajatdhasmana@gmail.com","username":"whoami-rajat"},"change_message_id":"6cbf33f5b75e6e55c5bed5456711a1fc0b0d5aff","unresolved":true,"context_lines":[{"line_number":202,"context_line":"    def _same_filesystem(self, path1, path2):"},{"line_number":203,"context_line":"        \"\"\"Return true if the two paths are in the same GPFS file system.\"\"\""},{"line_number":204,"context_line":"        try:"},{"line_number":205,"context_line":"            (out, err) \u003d self.gpfs_execute(\u0027stat\u0027, \u0027-f\u0027, \u0027-c\u0027, \u0027\"%i\"\u0027,"},{"line_number":206,"context_line":"                                           path1, path2)"},{"line_number":207,"context_line":"            lines \u003d out.splitlines()"},{"line_number":208,"context_line":"            return lines[0] \u003d\u003d lines[1]"},{"line_number":209,"context_line":"        except processutils.ProcessExecutionError as exc:"}],"source_content_type":"text/x-python","patch_set":4,"id":"d8002159_f4e81c9a","line":206,"range":{"start_line":205,"start_character":12,"end_line":206,"end_character":56},"updated":"2021-10-27 06:00:38.000000000","message":"I\u0027m not sure how replacing the lstat python method to stat command addresses the issue? If the pathnames are symlinks then we can use os.stat rather than the manual executing here\nbut i also see that this is only executed once whereas os.stat needs to be called twice","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":27615,"name":"Rajat Dhasmana","email":"rajatdhasmana@gmail.com","username":"whoami-rajat"},"change_message_id":"628b37d4c872df8b3fc783d186d1ded0c5a371d6","unresolved":true,"context_lines":[{"line_number":202,"context_line":"    def _same_filesystem(self, path1, path2):"},{"line_number":203,"context_line":"        \"\"\"Return true if the two paths are in the same GPFS file system.\"\"\""},{"line_number":204,"context_line":"        try:"},{"line_number":205,"context_line":"            (out, err) \u003d self.gpfs_execute(\u0027stat\u0027, \u0027-f\u0027, \u0027-c\u0027, \u0027\"%i\"\u0027,"},{"line_number":206,"context_line":"                                           path1, path2)"},{"line_number":207,"context_line":"            lines \u003d out.splitlines()"},{"line_number":208,"context_line":"            return lines[0] \u003d\u003d lines[1]"},{"line_number":209,"context_line":"        except processutils.ProcessExecutionError as exc:"}],"source_content_type":"text/x-python","patch_set":4,"id":"ffad7ff2_4af87cd8","line":206,"range":{"start_line":205,"start_character":12,"end_line":206,"end_character":56},"in_reply_to":"d208d086_9120b380","updated":"2021-10-28 14:33:42.000000000","message":"Ah, makes sense, i see the gpfs_execute reference is pointing to different methods in different classes","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"cc6e96b031687a2d4adca2ca2f5fe6d524616089","unresolved":true,"context_lines":[{"line_number":202,"context_line":"    def _same_filesystem(self, path1, path2):"},{"line_number":203,"context_line":"        \"\"\"Return true if the two paths are in the same GPFS file system.\"\"\""},{"line_number":204,"context_line":"        try:"},{"line_number":205,"context_line":"            (out, err) \u003d self.gpfs_execute(\u0027stat\u0027, \u0027-f\u0027, \u0027-c\u0027, \u0027\"%i\"\u0027,"},{"line_number":206,"context_line":"                                           path1, path2)"},{"line_number":207,"context_line":"            lines \u003d out.splitlines()"},{"line_number":208,"context_line":"            return lines[0] \u003d\u003d lines[1]"},{"line_number":209,"context_line":"        except processutils.ProcessExecutionError as exc:"}],"source_content_type":"text/x-python","patch_set":4,"id":"d208d086_9120b380","line":206,"range":{"start_line":205,"start_character":12,"end_line":206,"end_character":56},"in_reply_to":"d8002159_f4e81c9a","updated":"2021-10-27 12:59:10.000000000","message":"Rajat, my understanding is this driver supports multiple modes, some of which access the GPFS file system running on the same node where c-vol runs, and others where the GPFS file system is remote. I think the key is the code is using gpfs_execute() to ensure the check executes wherever the GPFS file system is located","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"26ef7911ee1532c1b92ce8c6c475f899d29feea7","unresolved":false,"context_lines":[{"line_number":202,"context_line":"    def _same_filesystem(self, path1, path2):"},{"line_number":203,"context_line":"        \"\"\"Return true if the two paths are in the same GPFS file system.\"\"\""},{"line_number":204,"context_line":"        try:"},{"line_number":205,"context_line":"            (out, err) \u003d self.gpfs_execute(\u0027stat\u0027, \u0027-f\u0027, \u0027-c\u0027, \u0027\"%i\"\u0027,"},{"line_number":206,"context_line":"                                           path1, path2)"},{"line_number":207,"context_line":"            lines \u003d out.splitlines()"},{"line_number":208,"context_line":"            return lines[0] \u003d\u003d lines[1]"},{"line_number":209,"context_line":"        except processutils.ProcessExecutionError as exc:"}],"source_content_type":"text/x-python","patch_set":4,"id":"c73b4987_5dc2ec12","line":206,"range":{"start_line":205,"start_character":12,"end_line":206,"end_character":56},"in_reply_to":"ffad7ff2_4af87cd8","updated":"2021-11-02 15:43:22.000000000","message":"Clarification given. Hence marked as Done.","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"4489fe9b6db69d32e9c09e62eb3a5a6fca569a67","unresolved":true,"context_lines":[{"line_number":917,"context_line":""},{"line_number":918,"context_line":"        data \u003d image_utils.qemu_img_info(image_path)"},{"line_number":919,"context_line":""},{"line_number":920,"context_line":"        if data.file_format \u003d\u003d \u0027raw\u0027:"},{"line_number":921,"context_line":"            vol_path \u003d self._get_volume_path(volume)"},{"line_number":922,"context_line":"        else:"},{"line_number":923,"context_line":"            vol_path \u003d self.local_path(volume)"}],"source_content_type":"text/x-python","patch_set":4,"id":"f7af24fd_0738a97a","line":920,"updated":"2021-10-25 17:15:57.000000000","message":"This conditional is immediately repeated on L927, so this one is redundant.\n- Move L921 so it follows L927\n- Move L923 so it follows L944\n\nIt took me a while to understand what\u0027s going on, so it would also help if a comment explained the vol_path is either local or remote based on whether the copy/clone operation takes place locally or on the GPFS backend.","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"26ef7911ee1532c1b92ce8c6c475f899d29feea7","unresolved":false,"context_lines":[{"line_number":917,"context_line":""},{"line_number":918,"context_line":"        data \u003d image_utils.qemu_img_info(image_path)"},{"line_number":919,"context_line":""},{"line_number":920,"context_line":"        if data.file_format \u003d\u003d \u0027raw\u0027:"},{"line_number":921,"context_line":"            vol_path \u003d self._get_volume_path(volume)"},{"line_number":922,"context_line":"        else:"},{"line_number":923,"context_line":"            vol_path \u003d self.local_path(volume)"}],"source_content_type":"text/x-python","patch_set":4,"id":"e2ffad76_9da42f02","line":920,"in_reply_to":"f7af24fd_0738a97a","updated":"2021-11-02 15:43:22.000000000","message":"Addressed in the latest patchset. Done","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":27615,"name":"Rajat Dhasmana","email":"rajatdhasmana@gmail.com","username":"whoami-rajat"},"change_message_id":"6cbf33f5b75e6e55c5bed5456711a1fc0b0d5aff","unresolved":true,"context_lines":[{"line_number":1598,"context_line":""},{"line_number":1599,"context_line":"    def local_path(self, volume):"},{"line_number":1600,"context_line":"        \"\"\"Returns the local path for the specified volume.\"\"\""},{"line_number":1601,"context_line":"        remotefs_share \u003d self._find_share(volume)"},{"line_number":1602,"context_line":"        base_local_path \u003d self._get_mount_point_for_share(remotefs_share)"},{"line_number":1603,"context_line":""},{"line_number":1604,"context_line":"        # Check if the volume is part of a consistency group and return"}],"source_content_type":"text/x-python","patch_set":4,"id":"4e8346c6_2b37c2f8","line":1601,"range":{"start_line":1601,"start_character":25,"end_line":1601,"end_character":49},"updated":"2021-10-27 06:00:38.000000000","message":"i think this is only needed when called from _clone_image (or in a situation where new volume is created) else the provider location in the volume should work fine.\nMy concern is this call will degrade performance of existing operations using this method for existing volumes.\nIt would be better to pass a is_clone_image (or is_new_vol) parameter having default\u003dFalse which will call NFS driver\u0027s _find_share method if true else we can rely on provider location.","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"3b6f3cfd2278fa7ed2fd3cceb20b16ba3bd10290","unresolved":true,"context_lines":[{"line_number":1598,"context_line":""},{"line_number":1599,"context_line":"    def local_path(self, volume):"},{"line_number":1600,"context_line":"        \"\"\"Returns the local path for the specified volume.\"\"\""},{"line_number":1601,"context_line":"        remotefs_share \u003d self._find_share(volume)"},{"line_number":1602,"context_line":"        base_local_path \u003d self._get_mount_point_for_share(remotefs_share)"},{"line_number":1603,"context_line":""},{"line_number":1604,"context_line":"        # Check if the volume is part of a consistency group and return"}],"source_content_type":"text/x-python","patch_set":4,"id":"624bb546_c54ead74","line":1601,"range":{"start_line":1601,"start_character":25,"end_line":1601,"end_character":49},"in_reply_to":"10ee2b9b_7b3845a9","updated":"2021-11-02 23:15:34.000000000","message":"Hi Rajat, Can you share your comment on this? Also, can you share your IRC ID for further communication?","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"26ef7911ee1532c1b92ce8c6c475f899d29feea7","unresolved":true,"context_lines":[{"line_number":1598,"context_line":""},{"line_number":1599,"context_line":"    def local_path(self, volume):"},{"line_number":1600,"context_line":"        \"\"\"Returns the local path for the specified volume.\"\"\""},{"line_number":1601,"context_line":"        remotefs_share \u003d self._find_share(volume)"},{"line_number":1602,"context_line":"        base_local_path \u003d self._get_mount_point_for_share(remotefs_share)"},{"line_number":1603,"context_line":""},{"line_number":1604,"context_line":"        # Check if the volume is part of a consistency group and return"}],"source_content_type":"text/x-python","patch_set":4,"id":"10ee2b9b_7b3845a9","line":1601,"range":{"start_line":1601,"start_character":25,"end_line":1601,"end_character":49},"in_reply_to":"4e8346c6_2b37c2f8","updated":"2021-11-02 15:43:22.000000000","message":"Apart from new volume, there are other situations where provider location returns NULL(for example, instance where volume get into ERROR state where it returns provider location as NULL). Hence, it has been handled using drivers _find_share method.","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"}],"releasenotes/notes/bug-1947123-87df5e176a86d73a.yaml":[{"author":{"_account_id":27615,"name":"Rajat Dhasmana","email":"rajatdhasmana@gmail.com","username":"whoami-rajat"},"change_message_id":"6cbf33f5b75e6e55c5bed5456711a1fc0b0d5aff","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":4,"id":"c19c27b3_17e54609","line":7,"range":{"start_line":1,"start_character":0,"end_line":7,"end_character":0},"updated":"2021-10-27 06:00:38.000000000","message":"I think we can merge both fixes columns in one releasenote with a different name as of bug number like fix-gpfs-nfs-cow.","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"26ef7911ee1532c1b92ce8c6c475f899d29feea7","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":4,"id":"8f653cf7_2a86d12e","line":7,"range":{"start_line":1,"start_character":0,"end_line":7,"end_character":0},"in_reply_to":"c19c27b3_17e54609","updated":"2021-11-02 15:43:22.000000000","message":"Addressed in the latest patchset. Done","commit_id":"bbb127f4b6faf500d0cf32460047531f91714555"}],"releasenotes/notes/bug-1947134-64bb5565efb71cca.yaml":[{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"cd30efad9a00b744fd9cb1689b5d6ff07da60926","unresolved":true,"context_lines":[{"line_number":8,"context_line":"    is set to copy_on_write."},{"line_number":9,"context_line":"  - |"},{"line_number":10,"context_line":"    `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed"},{"line_number":11,"context_line":"    the volume deletion issue when volume provider location provides null"},{"line_number":12,"context_line":"    for volume in error state."}],"source_content_type":"text/x-yaml","patch_set":7,"id":"b2a15b06_4503aee7","line":11,"updated":"2021-11-01 13:52:34.000000000","message":"I don\u0027t see this fix in the code, and LP bug #1947123 describes a different problem.\n\nIt would also help to add a prefix like \"IBM GPFS driver `Bug #...\" because it clarifies that this fix is for a specific driver. See [1] for how other driver fixes appear in the release notes (the GPFS bug fix notes don\u0027t follow that pattern).\n\n[1] https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_000/813984/7/check/build-openstack-releasenotes/0009322/docs/unreleased.html","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"aa866f49e1ada105d79b89a6626348774c3e792b","unresolved":true,"context_lines":[{"line_number":8,"context_line":"    is set to copy_on_write."},{"line_number":9,"context_line":"  - |"},{"line_number":10,"context_line":"    `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed"},{"line_number":11,"context_line":"    the volume deletion issue when volume provider location provides null"},{"line_number":12,"context_line":"    for volume in error state."}],"source_content_type":"text/x-yaml","patch_set":7,"id":"d2bfb966_9f2ba2c5","line":11,"in_reply_to":"14e10b47_4dee3d3f","updated":"2021-11-02 18:00:20.000000000","message":"Oh, now I see how the change also fixes the delete_volume issue! The changes to fix the local_path prevent the crash reported in the Red Hat bug, and other code [1],[2] gracefully handle situations where the volume wasn\u0027t actually created.\n\n[1] https://review.opendev.org/c/openstack/cinder/+/813984/7/cinder/volume/drivers/ibm/gpfs.py#636\n[2] https://review.opendev.org/c/openstack/cinder/+/813984/7/cinder/volume/drivers/ibm/gpfs.py#641\n\nSo I think the only issue is with the release note for LP #1947123. I like what you wrote:\n\n    IBM GPFS driver `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed\n    the volume creation issue in GPFS NFS driver when gpfs_images_share_mode\n    is set to copy_on_write.\n\nThis is fine, and I would just drop any reference to fixing the delete_volume problem. That issue isn\u0027t mentioned in the LP bug, which is why I got confused. Cleaning up the Red Hat bug is a separate issue.","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"fd67371dc34eac07ff1c2c2712a43777fe319875","unresolved":true,"context_lines":[{"line_number":8,"context_line":"    is set to copy_on_write."},{"line_number":9,"context_line":"  - |"},{"line_number":10,"context_line":"    `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed"},{"line_number":11,"context_line":"    the volume deletion issue when volume provider location provides null"},{"line_number":12,"context_line":"    for volume in error state."}],"source_content_type":"text/x-yaml","patch_set":7,"id":"c742cef9_390679b0","line":11,"in_reply_to":"58bb1388_964b42fe","updated":"2021-11-02 15:58:34.000000000","message":"I believe using self._find_share() does NOT solve the volume deletion problem. You can still get a volume in the error state in situations where it hasn\u0027t been created on the backend, in which case I suspect self._find_share() will still return None. Please review the comments in this Red Hat bug https://bugzilla.redhat.com/show_bug.cgi?id\u003d1985208.","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"},{"author":{"_account_id":21129,"name":"Alan Bishop","email":"abishopsweng@gmail.com","username":"ASBishop","status":"ex Red Hat"},"change_message_id":"c932f81b3f7f4850919acad8d1e2f0628621c964","unresolved":true,"context_lines":[{"line_number":8,"context_line":"    is set to copy_on_write."},{"line_number":9,"context_line":"  - |"},{"line_number":10,"context_line":"    `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed"},{"line_number":11,"context_line":"    the volume deletion issue when volume provider location provides null"},{"line_number":12,"context_line":"    for volume in error state."}],"source_content_type":"text/x-yaml","patch_set":7,"id":"a975fc82_528b6cfc","line":11,"in_reply_to":"66677585_4128d865","updated":"2021-11-02 18:51:12.000000000","message":"That all looks fine to me (obviously each reviewer has their own opinion).","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"f7aa10158108c318ab02768579ebf81562628ee3","unresolved":false,"context_lines":[{"line_number":8,"context_line":"    is set to copy_on_write."},{"line_number":9,"context_line":"  - |"},{"line_number":10,"context_line":"    `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed"},{"line_number":11,"context_line":"    the volume deletion issue when volume provider location provides null"},{"line_number":12,"context_line":"    for volume in error state."}],"source_content_type":"text/x-yaml","patch_set":7,"id":"acfd7521_90abb6fd","line":11,"in_reply_to":"9894da2b_1128be3a","updated":"2021-11-02 23:13:17.000000000","message":"Done","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"4bc2ad4f11ce3f44011b5165142a26be340cdcfe","unresolved":true,"context_lines":[{"line_number":8,"context_line":"    is set to copy_on_write."},{"line_number":9,"context_line":"  - |"},{"line_number":10,"context_line":"    `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed"},{"line_number":11,"context_line":"    the volume deletion issue when volume provider location provides null"},{"line_number":12,"context_line":"    for volume in error state."}],"source_content_type":"text/x-yaml","patch_set":7,"id":"9894da2b_1128be3a","line":11,"in_reply_to":"a975fc82_528b6cfc","updated":"2021-11-02 23:12:43.000000000","message":"Addressed in the latest patchset. Done","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"26ef7911ee1532c1b92ce8c6c475f899d29feea7","unresolved":true,"context_lines":[{"line_number":8,"context_line":"    is set to copy_on_write."},{"line_number":9,"context_line":"  - |"},{"line_number":10,"context_line":"    `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed"},{"line_number":11,"context_line":"    the volume deletion issue when volume provider location provides null"},{"line_number":12,"context_line":"    for volume in error state."}],"source_content_type":"text/x-yaml","patch_set":7,"id":"58bb1388_964b42fe","line":11,"in_reply_to":"b2a15b06_4503aee7","updated":"2021-11-02 15:43:22.000000000","message":"1. Replacing release notes content as given below. Is that okay?\n\n---\nfixes:\n\n  - |\n    IBM GPFS driver `Bug #1947134 \u003chttps://bugs.launchpad.net/cinder/+bug/1947134\u003e`_: Fixed\n    the initialization of GPFS NFS driver when gpfs_images_share_mode\n    is set to copy_on_write by correcting _same_filesystem functionality.\n  - |\n    IBM GPFS driver `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed\n    the volume creation issue in GPFS NFS driver when gpfs_images_share_mode\n    is set to copy_on_write.\n    the volume deletion issue when volume provider location provides null\n    for volume in error state.\n\n2. Wrt volume deletion issue(provider location \u003d NULL), this issue is fixed by line 1600\"remotefs_share \u003d self._find_share(volume)\" - i.e. volume path is always available with _find_share() method in local_path()\n def local_path(self, volume):\n        \"\"\"Returns the local path for the specified volume.\"\"\"\n        remotefs_share \u003d self._find_share(volume)\n\nThe following scenarios are happening with our driver.\n1. In copy mode, volume physical file is not created and provider location return as NULL\n2. In copy on write mode, volume physical file is created and provider location return as NULL\nThe following line( remotefs_share \u003d self._find_share(volume)) handles both above given scenario(i.e. It deletes if volume file is physically present during this error scenario. If I use volume provider_location NULL check as a fix then it does not delete the physical file in COW mode)\n\nThis fix(remotefs_share \u003d self._find_share(volume)) handles both volume deletion issue(for volume in error state) and volume creation issue in cow mode.","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"a22537bb99d557eff465c19e11cecf95c988f07e","unresolved":true,"context_lines":[{"line_number":8,"context_line":"    is set to copy_on_write."},{"line_number":9,"context_line":"  - |"},{"line_number":10,"context_line":"    `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed"},{"line_number":11,"context_line":"    the volume deletion issue when volume provider location provides null"},{"line_number":12,"context_line":"    for volume in error state."}],"source_content_type":"text/x-yaml","patch_set":7,"id":"14e10b47_4dee3d3f","line":11,"in_reply_to":"c742cef9_390679b0","updated":"2021-11-02 17:19:29.000000000","message":"Thanks for the response. Yes, I have experimented/simulated the scenario where volume in error state and volume file did not create in the backend where provider location returns NULL and self._find_share()  returns proper share information. Please let us know if we need to cover any other scenarios in testing.\n\nSample logs given below.\n\n2021-11-02 17:13:35.127 91 ERROR oslo_messaging.rpc.server   File \"/usr/lib/python3.6/site-packages/cinder/image/image_utils.py\", line 742, in check_virtual_size\n2021-11-02 17:13:35.127 91 ERROR oslo_messaging.rpc.server     reason\u003dreason)\n2021-11-02 17:13:35.127 91 ERROR oslo_messaging.rpc.server cinder.exception.ImageUnacceptable: Image f78a33da-6325-47e4-ac63-d8bb83a93ee8 is unacceptable: Image virtual size is 10GB and doesn\u0027t fit in a volume of size 2GB.\n2021-11-02 17:13:35.127 91 ERROR oslo_messaging.rpc.server \n\n2021-11-02 17:14:26.030 91 ERROR cinder.volume.drivers.ibm.gpfs [req-f5cf7c9b-928b-4f4e-98e4-bc70a8ee2400 7f77a6426b854960995f46195f94751d 910f4c29b4ed4eb8857b0d874b7197ec - default default] remotefs_share: 192.168.24.111:/ibm/fs1/openstack/mode3/cinder/volumes\n2021-11-02 17:14:26.031 91 ERROR cinder.volume.drivers.ibm.gpfs [req-f5cf7c9b-928b-4f4e-98e4-bc70a8ee2400 7f77a6426b854960995f46195f94751d 910f4c29b4ed4eb8857b0d874b7197ec - default default] provider_location: None\n2021-11-02 17:14:26.071 91 WARNING py.warnings [req-f5cf7c9b-928b-4f4e-98e4-bc70a8ee2400 7f77a6426b854960995f46195f94751d 910f4c29b4ed4eb8857b0d874b7197ec - default default] /usr/lib64/python3.6/site-packages/sqlalchemy/orm/evaluator.py:99: SAWarning: Evaluating non-mapped column expression \u0027updated_at\u0027 onto ORM instances; this is a deprecated use case.  Please make use of the actual mapped columns in ORM-evaluated UPDATE / DELETE expressions.\n  \"UPDATE / DELETE expressions.\" % clause\n\n2021-11-02 17:14:26.141 91 INFO cinder.volume.manager [req-f5cf7c9b-928b-4f4e-98e4-bc70a8ee2400 7f77a6426b854960995f46195f94751d 910f4c29b4ed4eb8857b0d874b7197ec - default default] Deleted volume successfully.","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"},{"author":{"_account_id":34167,"name":"rajadeepan","email":"rajadeepan.ramar@ibm.com","username":"rajadeepan"},"change_message_id":"154fd3c3f3bb6db16dd3f67b7e59e08f118c3c24","unresolved":true,"context_lines":[{"line_number":8,"context_line":"    is set to copy_on_write."},{"line_number":9,"context_line":"  - |"},{"line_number":10,"context_line":"    `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed"},{"line_number":11,"context_line":"    the volume deletion issue when volume provider location provides null"},{"line_number":12,"context_line":"    for volume in error state."}],"source_content_type":"text/x-yaml","patch_set":7,"id":"66677585_4128d865","line":11,"in_reply_to":"d2bfb966_9f2ba2c5","updated":"2021-11-02 18:46:56.000000000","message":"Thanks for the review. I will rename the release notes file name as \"bug-gpfs-fix-nfs-cow.yaml\" with content as given below. Is that fine(particularly release notes file name)?\n\n---\nfixes:\n\n  - |\n    IBM GPFS driver `Bug #1947134 \u003chttps://bugs.launchpad.net/cinder/+bug/1947134\u003e`_: Fixed\n    the initialization of GPFS NFS driver when gpfs_images_share_mode\n    is set to copy_on_write by correcting _same_filesystem functionality.\n  - |\n    IBM GPFS driver `Bug #1947123 \u003chttps://bugs.launchpad.net/cinder/+bug/1947123\u003e`_: Fixed\n    the volume creation issue in GPFS NFS driver when gpfs_images_share_mode\n    is set to copy_on_write.","commit_id":"5f7936431c3028762e3ab8728404850bca3acc52"}]}
