)]}'
{"/COMMIT_MSG":[{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"c6169437edc84638c72a8ffb1c0f044814150d26","unresolved":true,"context_lines":[{"line_number":14,"context_line":"both backends equally, which will result in one backend getting all of"},{"line_number":15,"context_line":"the share creation requests and the other backend getting none."},{"line_number":16,"context_line":""},{"line_number":17,"context_line":"To fix this have the driver report the allocated_capacity_gb along with"},{"line_number":18,"context_line":"the provisioned_capacity_gb, which is the sum of the shares assigned to"},{"line_number":19,"context_line":"that backend.  Then if you enable thin_provisioning and thin_logic the"},{"line_number":20,"context_line":"capacity weigher will use the least allocated cephfs backend as a higher"},{"line_number":21,"context_line":"candidate for share creates.  Which will result in share creations being"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":10,"id":"d6bec6f9_95013c5f","line":18,"range":{"start_line":17,"start_character":61,"end_line":18,"end_character":27},"updated":"2024-09-12 18:26:00.000000000","message":"As we discussed earlier,\n\nThe CephFS driver looking at the sizes of all subvolumes in a FS to determine an \"allocated_capacity_gb\" is fine, and correct.. however, that\u0027s not the same as provisioned_capacity_gb always. \n\n`provisioned_capacity_gb` tracks what was provisioned by an instance of the manila driver. It\u0027s ideally something that Manila knows about/can calculate. By setting it the same as `allocated_capacity_gb`, and enabling `thin_provisioning` at the same time, you\u0027re effectively controlling oversubscription through the CephFS share driver. Without also allowing the configuration of, and reporting `max_over_subscription_ratio`, this can cause some pain.. \n\n\nCan we please separate the concerns within this patch?\n\nPlease report \u0027allocated_capacity_gb\u0027 the way you\u0027re estimating it; it allows the scheduler to have this information so goodness_function and filter_function can be used to control provisioning. This bugfix can be backported to any active stable branches.\n\nIn a different change, switch the driver to report thin_provisioning\u003dTrue. \nThis enables the storage\u0027s oversubscription to be controlled through manila; make sure to read and report the value of max_oversubscription_ratio so that the scheduler can calculate the appropriate provisioning ratio. I would not suggest backporting this change, you can break existing deployments that have share type extra specifications, and tunings to explicitly look for the scheduler parameter. I\u0027d suggest landing this change land alongside haixin\u0027s change; so you\u0027re deliberately falling back to allocated_capacity_gb when provisioned_capacity_gb is unknown.\n\n\nhttps://review.opendev.org/c/openstack/manila/+/838672/6/manila/scheduler/host_manager.py#468","commit_id":"2224fc0edd309b9ef964f3958cf0159a1d22901b"},{"author":{"_account_id":30407,"name":"haixin","email":"haixin_haixin@qq.com","username":"haixin"},"change_message_id":"5edf0de5776f947fc729861aa6e4a99ad05ad6ec","unresolved":true,"context_lines":[{"line_number":14,"context_line":"both backends equally, which will result in one backend getting all of"},{"line_number":15,"context_line":"the share creation requests and the other backend getting none."},{"line_number":16,"context_line":""},{"line_number":17,"context_line":"To fix this have the driver report the allocated_capacity_gb along with"},{"line_number":18,"context_line":"the provisioned_capacity_gb, which is the sum of the shares assigned to"},{"line_number":19,"context_line":"that backend.  Then if you enable thin_provisioning and thin_logic the"},{"line_number":20,"context_line":"capacity weigher will use the least allocated cephfs backend as a higher"},{"line_number":21,"context_line":"candidate for share creates.  Which will result in share creations being"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":10,"id":"948b66c1_20a5a4fd","line":18,"range":{"start_line":17,"start_character":61,"end_line":18,"end_character":27},"in_reply_to":"5b24c9a4_17ee8bb3","updated":"2025-01-10 03:25:48.000000000","message":"Currently, the cephfs driver reports only allocated_capacity_gb, which does not conflict with my patch. In my patch manila will calculate allocated_capacity_gb by itself, if the driver does not report provisioned_capacity_gb, allocated_capacity_gb will be used by default. I agree with this patch merge.","commit_id":"2224fc0edd309b9ef964f3958cf0159a1d22901b"},{"author":{"_account_id":30407,"name":"haixin","email":"haixin_haixin@qq.com","username":"haixin"},"change_message_id":"5d9bea11283402545a8c78dd81bda60455329a93","unresolved":true,"context_lines":[{"line_number":14,"context_line":"both backends equally, which will result in one backend getting all of"},{"line_number":15,"context_line":"the share creation requests and the other backend getting none."},{"line_number":16,"context_line":""},{"line_number":17,"context_line":"To fix this have the driver report the allocated_capacity_gb along with"},{"line_number":18,"context_line":"the provisioned_capacity_gb, which is the sum of the shares assigned to"},{"line_number":19,"context_line":"that backend.  Then if you enable thin_provisioning and thin_logic the"},{"line_number":20,"context_line":"capacity weigher will use the least allocated cephfs backend as a higher"},{"line_number":21,"context_line":"candidate for share creates.  Which will result in share creations being"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":10,"id":"f74b1045_f077a0d8","line":18,"range":{"start_line":17,"start_character":61,"end_line":18,"end_character":27},"in_reply_to":"c71efae8_01ee4fd3","updated":"2024-09-23 06:50:40.000000000","message":"https://opendev.org/openstack/manila/src/commit/dca04e2270a745e73226adbdf80f740f002bd133/manila/scheduler/host_manager.py#L472\n\nprovisioned_capacity_gb is equal to the sum of allocated_capacity_gb and\nshares_gb_created_manually (or another manila) on the same storage pool.\n\nprovisioned_capacity_gb \u003d Total capacity allocated by all shares (include\nshares created manually or other manila) in the storage pool.\n\nallocated_capacity_gb \u003d Total capacity allocated by all shares created by\ncurrent manila.\n\nso provisioned_capacity_gb \u003e\u003d allocated_capacity_gb\n\nManila is considered to be monopolizing the storage back-end if provisioned_capacity_gb \u003d\u003d allocated_capacity_gb. \n\nIn my opinion, the driver layer should report provisioned_capacity_gb. However, allocated_capacity_gb cannot be reported. This is because the back end does not know which subvolumes were created and used by the current manila and which subvolumes were created manually or used by other manila services.\n\nAs with the changes in my patch, manila calculates the allocated_capacity_gb for each host on its own, without backend reporting. Because the allocated_capacity_gb statistics from the database are completely accurate. As mentioned above, the allocated_capacity_gb reported by the backend is inaccurate.\n\nWe only need the backend to report provisioned_capacity_gb because manila does not know if there are other manually created share instances on this backend.\n\nFinally, i will take some time to update that patch.\n838672: manila oversubscription enhancements | https://review.opendev.org/c/openstack/manila/+/838672","commit_id":"2224fc0edd309b9ef964f3958cf0159a1d22901b"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"5da32dae5d9ef04bffe87ddad103d0633d5fbe51","unresolved":true,"context_lines":[{"line_number":14,"context_line":"both backends equally, which will result in one backend getting all of"},{"line_number":15,"context_line":"the share creation requests and the other backend getting none."},{"line_number":16,"context_line":""},{"line_number":17,"context_line":"To fix this have the driver report the allocated_capacity_gb along with"},{"line_number":18,"context_line":"the provisioned_capacity_gb, which is the sum of the shares assigned to"},{"line_number":19,"context_line":"that backend.  Then if you enable thin_provisioning and thin_logic the"},{"line_number":20,"context_line":"capacity weigher will use the least allocated cephfs backend as a higher"},{"line_number":21,"context_line":"candidate for share creates.  Which will result in share creations being"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":10,"id":"c71efae8_01ee4fd3","line":18,"range":{"start_line":17,"start_character":61,"end_line":18,"end_character":27},"in_reply_to":"d6bec6f9_95013c5f","updated":"2024-09-12 18:28:26.000000000","message":"On the second change, as an alternative, if you make this thin_provisioning behavior a configuration option that defaults to the current state, it can be backportable","commit_id":"2224fc0edd309b9ef964f3958cf0159a1d22901b"},{"author":{"_account_id":29632,"name":"Carlos Eduardo","email":"ces.eduardo98@gmail.com","username":"silvacarlos"},"change_message_id":"ae7716132741cffd39fd544f847bfa371b67173d","unresolved":true,"context_lines":[{"line_number":14,"context_line":"both backends equally, which will result in one backend getting all of"},{"line_number":15,"context_line":"the share creation requests and the other backend getting none."},{"line_number":16,"context_line":""},{"line_number":17,"context_line":"To fix this have the driver report the allocated_capacity_gb along with"},{"line_number":18,"context_line":"the provisioned_capacity_gb, which is the sum of the shares assigned to"},{"line_number":19,"context_line":"that backend.  Then if you enable thin_provisioning and thin_logic the"},{"line_number":20,"context_line":"capacity weigher will use the least allocated cephfs backend as a higher"},{"line_number":21,"context_line":"candidate for share creates.  Which will result in share creations being"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":10,"id":"5b24c9a4_17ee8bb3","line":18,"range":{"start_line":17,"start_character":61,"end_line":18,"end_character":27},"in_reply_to":"f74b1045_f077a0d8","updated":"2024-12-26 19:31:05.000000000","message":"I\u0027ve currently taken Goutham suggestion on the allocated capacity gb but I think we should get other opinions on your suggestion to have the driver only reporting provisioned capacity, Haixin! I think it is a good solution, so let\u0027s see if others agree with it too :)","commit_id":"2224fc0edd309b9ef964f3958cf0159a1d22901b"}],"/PATCHSET_LEVEL":[{"author":{"_account_id":29632,"name":"Carlos Eduardo","email":"ces.eduardo98@gmail.com","username":"silvacarlos"},"change_message_id":"c81e4d15771d8c46a19e85fac5a911b00f3c7012","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":6,"id":"e5d0da38_13280387","updated":"2024-02-20 21:21:08.000000000","message":"Hi, Kris! Thank you very much for reporting the issue and proposing the fix.\n\nApologies for taking some time to get back to you on this issue. This can be considered as a bug fix, so if we are willing to get this merged in the Caracal cycle, we have until RC1 for that.\n\nIn the meanwhile, I\u0027ve been putting some focus on the RFE to allow the filesystem name to be set in the share metadata [1], and this one should be merged by the feature freeze (next week), so that\u0027s why I\u0027m putting some more focus on it now, but I intend to get back to this bug soon.\n\n[1] https://review.opendev.org/c/openstack/manila/+/909175","commit_id":"6f1b90e664f3cc341229940c4a5c9f2f475fef43"},{"author":{"_account_id":29632,"name":"Carlos Eduardo","email":"ces.eduardo98@gmail.com","username":"silvacarlos"},"change_message_id":"06d2177dcaeb4265c23753a1508f27d177ba8ea5","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":8,"id":"f9b52d56_6537ccd7","updated":"2024-07-12 21:24:41.000000000","message":"Thank you for your thoughts and working on this change. I managed to put together some updates and figure out a way to cache the allocated capacity gb information in the driver. Please take a look at the most recent implementation and let me know your thoughts.\n\nI have also performed some tests: https://paste.opendev.org/show/bwRagTpFQdGNtLCfiUtE/","commit_id":"bf725ce90d68be963d8a190d41ec26c6594b0730"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"67a2855e199c328ae225f4607adf5f4cd17859cc","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":9,"id":"05d78118_8555d91d","updated":"2024-09-11 20:16:49.000000000","message":"A couple of minor issues; thanks Kris and Carlos!","commit_id":"6f3c6affdffebbd5a7fe429b14644ef6cc022570"},{"author":{"_account_id":29632,"name":"Carlos Eduardo","email":"ces.eduardo98@gmail.com","username":"silvacarlos"},"change_message_id":"5379f0222983a22b331d61abc0465dae606039b8","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":9,"id":"02ef40ab_370390f8","updated":"2024-09-11 21:16:52.000000000","message":"Thanks for the reviews, PTAL :)","commit_id":"6f3c6affdffebbd5a7fe429b14644ef6cc022570"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"c6169437edc84638c72a8ffb1c0f044814150d26","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":10,"id":"48e3680c_a3dceb4e","updated":"2024-09-12 18:26:00.000000000","message":"Hi Carlos, please see comments inline","commit_id":"2224fc0edd309b9ef964f3958cf0159a1d22901b"},{"author":{"_account_id":30407,"name":"haixin","email":"haixin_haixin@qq.com","username":"haixin"},"change_message_id":"5d9bea11283402545a8c78dd81bda60455329a93","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":10,"id":"7354dce2_e0fab3c4","updated":"2024-09-23 06:50:40.000000000","message":"I wrote some of my thoughts on drive report capacity_gb.","commit_id":"2224fc0edd309b9ef964f3958cf0159a1d22901b"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"9d6be66b7fa7a66dfd31fa197eff6b659761e07e","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":11,"id":"065bb0d2_5f7cc4fd","updated":"2025-01-10 20:20:32.000000000","message":"lgtm,thanks for pursuing this bug fix","commit_id":"d3d86bde80ba287c616af76a58d9f840cac0ce4a"},{"author":{"_account_id":30407,"name":"haixin","email":"haixin_haixin@qq.com","username":"haixin"},"change_message_id":"f2e0f9e27fa983e56268c1bea000af9bd1f7aa92","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":11,"id":"68fac9a1_969a5958","updated":"2025-01-10 03:26:08.000000000","message":"thanks, Carlos.","commit_id":"d3d86bde80ba287c616af76a58d9f840cac0ce4a"}],"manila/share/drivers/cephfs/driver.py":[{"author":{"_account_id":29632,"name":"Carlos Eduardo","email":"ces.eduardo98@gmail.com","username":"silvacarlos"},"change_message_id":"c81e4d15771d8c46a19e85fac5a911b00f3c7012","unresolved":true,"context_lines":[{"line_number":287,"context_line":"        free_capacity_gb \u003d round(stats[\u0027kb_avail\u0027] / units.Mi, 2)"},{"line_number":288,"context_line":"        allocated_capacity_gb \u003d 0"},{"line_number":289,"context_line":""},{"line_number":290,"context_line":"        argdict \u003d {\"vol_name\": self.volname}"},{"line_number":291,"context_line":"        subvolumes \u003d rados_command("},{"line_number":292,"context_line":"            self.rados_client, \"fs subvolume ls\", argdict, json_obj\u003dTrue)"},{"line_number":293,"context_line":"        for sub_vol in subvolumes:"},{"line_number":294,"context_line":"            argdict \u003d {\"vol_name\": self.volname, \"sub_name\": sub_vol[\"name\"]}"},{"line_number":295,"context_line":"            sub_info \u003d rados_command("},{"line_number":296,"context_line":"                self.rados_client, \"fs subvolume info\", argdict, json_obj\u003dTrue)"},{"line_number":297,"context_line":"            size \u003d sub_info.get(\u0027bytes_quota\u0027, 0)"},{"line_number":298,"context_line":"            if size \u003d\u003d \"infinite\":"},{"line_number":299,"context_line":"                # If we have a share that has infinite quota, we should not"}],"source_content_type":"text/x-python","patch_set":6,"id":"e9c22cbf_9e4f2323","line":296,"range":{"start_line":290,"start_character":8,"end_line":296,"end_character":79},"updated":"2024-02-20 21:21:08.000000000","message":"I am mainly concerned about doing this every single time we try to get the share stats... There might be hundreds or thousands of Manila shares, and sometimes the ceph cluster can be busy and it could take a while... I\u0027m looking into different approaches to suggest :)","commit_id":"6f1b90e664f3cc341229940c4a5c9f2f475fef43"},{"author":{"_account_id":29632,"name":"Carlos Eduardo","email":"ces.eduardo98@gmail.com","username":"silvacarlos"},"change_message_id":"5379f0222983a22b331d61abc0465dae606039b8","unresolved":false,"context_lines":[{"line_number":287,"context_line":"        free_capacity_gb \u003d round(stats[\u0027kb_avail\u0027] / units.Mi, 2)"},{"line_number":288,"context_line":"        allocated_capacity_gb \u003d 0"},{"line_number":289,"context_line":""},{"line_number":290,"context_line":"        argdict \u003d {\"vol_name\": self.volname}"},{"line_number":291,"context_line":"        subvolumes \u003d rados_command("},{"line_number":292,"context_line":"            self.rados_client, \"fs subvolume ls\", argdict, json_obj\u003dTrue)"},{"line_number":293,"context_line":"        for sub_vol in subvolumes:"},{"line_number":294,"context_line":"            argdict \u003d {\"vol_name\": self.volname, \"sub_name\": sub_vol[\"name\"]}"},{"line_number":295,"context_line":"            sub_info \u003d rados_command("},{"line_number":296,"context_line":"                self.rados_client, \"fs subvolume info\", argdict, json_obj\u003dTrue)"},{"line_number":297,"context_line":"            size \u003d sub_info.get(\u0027bytes_quota\u0027, 0)"},{"line_number":298,"context_line":"            if size \u003d\u003d \"infinite\":"},{"line_number":299,"context_line":"                # If we have a share that has infinite quota, we should not"}],"source_content_type":"text/x-python","patch_set":6,"id":"bff10fee_d1f379a9","line":296,"range":{"start_line":290,"start_character":8,"end_line":296,"end_character":79},"in_reply_to":"5e6b5e47_19d6d33c","updated":"2024-09-11 21:16:52.000000000","message":"Done","commit_id":"6f1b90e664f3cc341229940c4a5c9f2f475fef43"},{"author":{"_account_id":9076,"name":"Kris Lindgren","email":"klindgren@godaddy.com","username":"klindgren"},"change_message_id":"d71fdb917ec5576be03e180fa64c98403f6a8856","unresolved":true,"context_lines":[{"line_number":287,"context_line":"        free_capacity_gb \u003d round(stats[\u0027kb_avail\u0027] / units.Mi, 2)"},{"line_number":288,"context_line":"        allocated_capacity_gb \u003d 0"},{"line_number":289,"context_line":""},{"line_number":290,"context_line":"        argdict \u003d {\"vol_name\": self.volname}"},{"line_number":291,"context_line":"        subvolumes \u003d rados_command("},{"line_number":292,"context_line":"            self.rados_client, \"fs subvolume ls\", argdict, json_obj\u003dTrue)"},{"line_number":293,"context_line":"        for sub_vol in subvolumes:"},{"line_number":294,"context_line":"            argdict \u003d {\"vol_name\": self.volname, \"sub_name\": sub_vol[\"name\"]}"},{"line_number":295,"context_line":"            sub_info \u003d rados_command("},{"line_number":296,"context_line":"                self.rados_client, \"fs subvolume info\", argdict, json_obj\u003dTrue)"},{"line_number":297,"context_line":"            size \u003d sub_info.get(\u0027bytes_quota\u0027, 0)"},{"line_number":298,"context_line":"            if size \u003d\u003d \"infinite\":"},{"line_number":299,"context_line":"                # If we have a share that has infinite quota, we should not"}],"source_content_type":"text/x-python","patch_set":6,"id":"5e6b5e47_19d6d33c","line":296,"range":{"start_line":290,"start_character":8,"end_line":296,"end_character":79},"in_reply_to":"e9c22cbf_9e4f2323","updated":"2024-02-21 17:32:44.000000000","message":"Yes, that was my concern as well, and gathering this specific information every minute or every status run is totally unnecessary, it should only really change when Manila makes a change on the backend.\n\nMy other thought was that Manila already has this information in the DB, and imho this should be something that manila could track per backend (the number and size of shares allocated to a backend) and simply use that information in a specific scheduler.  Instead of relying on the backend to gather/sum/report this on a frequent cadence.  This would also allow the scheduler to filter/rank backends using the same filter independent of backend.","commit_id":"6f1b90e664f3cc341229940c4a5c9f2f475fef43"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"67a2855e199c328ae225f4607adf5f4cd17859cc","unresolved":true,"context_lines":[{"line_number":175,"context_line":"class AllocationCapacityCache(object):"},{"line_number":176,"context_line":"    \"\"\"AllocationCapacityCache for CephFS filesystems."},{"line_number":177,"context_line":""},{"line_number":178,"context_line":"    The cache validity is measured by a stop watch that is"},{"line_number":179,"context_line":"    not thread-safe."},{"line_number":180,"context_line":"    \"\"\""},{"line_number":181,"context_line":""},{"line_number":182,"context_line":"    def __init__(self, duration):"}],"source_content_type":"text/x-python","patch_set":9,"id":"0e0cb6c7_afb988ee","line":179,"range":{"start_line":178,"start_character":4,"end_line":179,"end_character":20},"updated":"2024-09-11 20:16:49.000000000","message":"I guess we need to note this in the release note as well..","commit_id":"6f3c6affdffebbd5a7fe429b14644ef6cc022570"},{"author":{"_account_id":29632,"name":"Carlos Eduardo","email":"ces.eduardo98@gmail.com","username":"silvacarlos"},"change_message_id":"5379f0222983a22b331d61abc0465dae606039b8","unresolved":false,"context_lines":[{"line_number":175,"context_line":"class AllocationCapacityCache(object):"},{"line_number":176,"context_line":"    \"\"\"AllocationCapacityCache for CephFS filesystems."},{"line_number":177,"context_line":""},{"line_number":178,"context_line":"    The cache validity is measured by a stop watch that is"},{"line_number":179,"context_line":"    not thread-safe."},{"line_number":180,"context_line":"    \"\"\""},{"line_number":181,"context_line":""},{"line_number":182,"context_line":"    def __init__(self, duration):"}],"source_content_type":"text/x-python","patch_set":9,"id":"51c58c70_d9a3975e","line":179,"range":{"start_line":178,"start_character":4,"end_line":179,"end_character":20},"in_reply_to":"0e0cb6c7_afb988ee","updated":"2024-09-11 21:16:52.000000000","message":"Done","commit_id":"6f3c6affdffebbd5a7fe429b14644ef6cc022570"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"67a2855e199c328ae225f4607adf5f4cd17859cc","unresolved":true,"context_lines":[{"line_number":366,"context_line":"                    \u0027total_capacity_gb\u0027: total_capacity_gb,"},{"line_number":367,"context_line":"                    \u0027free_capacity_gb\u0027: free_capacity_gb,"},{"line_number":368,"context_line":"                    \u0027allocated_capacity_gb\u0027: allocated_capacity_gb,"},{"line_number":369,"context_line":"                    \u0027provisioned_capacity_gb\u0027: allocated_capacity_gb,"},{"line_number":370,"context_line":"                    \u0027qos\u0027: \u0027False\u0027,"},{"line_number":371,"context_line":"                    \u0027reserved_percentage\u0027: self.configuration.safe_get("},{"line_number":372,"context_line":"                        \u0027reserved_share_percentage\u0027),"}],"source_content_type":"text/x-python","patch_set":9,"id":"b8f48865_00d74479","line":369,"range":{"start_line":369,"start_character":0,"end_line":369,"end_character":69},"updated":"2024-09-11 20:16:49.000000000","message":"let teh scheduler update this; \"provisioned_capacity_gb\" is technically what this manila instance has provisioend through the driver.. \n\nby calculating the size of all subvolumes in a given filesystem, you\u0027re really reporting allocated_capacity_gb","commit_id":"6f3c6affdffebbd5a7fe429b14644ef6cc022570"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"c6169437edc84638c72a8ffb1c0f044814150d26","unresolved":true,"context_lines":[{"line_number":366,"context_line":"                    \u0027total_capacity_gb\u0027: total_capacity_gb,"},{"line_number":367,"context_line":"                    \u0027free_capacity_gb\u0027: free_capacity_gb,"},{"line_number":368,"context_line":"                    \u0027allocated_capacity_gb\u0027: allocated_capacity_gb,"},{"line_number":369,"context_line":"                    \u0027provisioned_capacity_gb\u0027: allocated_capacity_gb,"},{"line_number":370,"context_line":"                    \u0027qos\u0027: \u0027False\u0027,"},{"line_number":371,"context_line":"                    \u0027reserved_percentage\u0027: self.configuration.safe_get("},{"line_number":372,"context_line":"                        \u0027reserved_share_percentage\u0027),"}],"source_content_type":"text/x-python","patch_set":9,"id":"79543aed_94d54a7d","line":369,"range":{"start_line":369,"start_character":0,"end_line":369,"end_character":69},"in_reply_to":"5bab58aa_6f4efe0e","updated":"2024-09-12 18:26:00.000000000","message":"this isn\u0027t done..","commit_id":"6f3c6affdffebbd5a7fe429b14644ef6cc022570"},{"author":{"_account_id":29632,"name":"Carlos Eduardo","email":"ces.eduardo98@gmail.com","username":"silvacarlos"},"change_message_id":"4ac495b9b2f06a0db2287d6c628028a079ecb74d","unresolved":false,"context_lines":[{"line_number":366,"context_line":"                    \u0027total_capacity_gb\u0027: total_capacity_gb,"},{"line_number":367,"context_line":"                    \u0027free_capacity_gb\u0027: free_capacity_gb,"},{"line_number":368,"context_line":"                    \u0027allocated_capacity_gb\u0027: allocated_capacity_gb,"},{"line_number":369,"context_line":"                    \u0027provisioned_capacity_gb\u0027: allocated_capacity_gb,"},{"line_number":370,"context_line":"                    \u0027qos\u0027: \u0027False\u0027,"},{"line_number":371,"context_line":"                    \u0027reserved_percentage\u0027: self.configuration.safe_get("},{"line_number":372,"context_line":"                        \u0027reserved_share_percentage\u0027),"}],"source_content_type":"text/x-python","patch_set":9,"id":"7361c03f_0866a01e","line":369,"range":{"start_line":369,"start_character":0,"end_line":369,"end_character":69},"in_reply_to":"79543aed_94d54a7d","updated":"2024-12-26 19:29:55.000000000","message":"removed the provisioned capacity gb from the list","commit_id":"6f3c6affdffebbd5a7fe429b14644ef6cc022570"},{"author":{"_account_id":29632,"name":"Carlos Eduardo","email":"ces.eduardo98@gmail.com","username":"silvacarlos"},"change_message_id":"5379f0222983a22b331d61abc0465dae606039b8","unresolved":false,"context_lines":[{"line_number":366,"context_line":"                    \u0027total_capacity_gb\u0027: total_capacity_gb,"},{"line_number":367,"context_line":"                    \u0027free_capacity_gb\u0027: free_capacity_gb,"},{"line_number":368,"context_line":"                    \u0027allocated_capacity_gb\u0027: allocated_capacity_gb,"},{"line_number":369,"context_line":"                    \u0027provisioned_capacity_gb\u0027: allocated_capacity_gb,"},{"line_number":370,"context_line":"                    \u0027qos\u0027: \u0027False\u0027,"},{"line_number":371,"context_line":"                    \u0027reserved_percentage\u0027: self.configuration.safe_get("},{"line_number":372,"context_line":"                        \u0027reserved_share_percentage\u0027),"}],"source_content_type":"text/x-python","patch_set":9,"id":"5bab58aa_6f4efe0e","line":369,"range":{"start_line":369,"start_character":0,"end_line":369,"end_character":69},"in_reply_to":"b8f48865_00d74479","updated":"2024-09-11 21:16:52.000000000","message":"Done","commit_id":"6f3c6affdffebbd5a7fe429b14644ef6cc022570"}]}
