)]}'
{".zuul.yaml":[{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"e6266804a2fbb6a2bb977b7a5ca7652c14d396cc","unresolved":true,"context_lines":[{"line_number":98,"context_line":"        CEPHADM_DEPLOY: True"},{"line_number":99,"context_line":"        MYSQL_REDUCE_MEMORY: True"},{"line_number":100,"context_line":"        REMOTE_CEPH: False"},{"line_number":101,"context_line":"      tempest_concurrency: 1"},{"line_number":102,"context_line":"    group-vars:"},{"line_number":103,"context_line":"      subnode:"},{"line_number":104,"context_line":"        devstack_localrc:"}],"source_content_type":"text/x-yaml","patch_set":41,"id":"7d260b5e_f000fc73","line":101,"range":{"start_line":101,"start_character":6,"end_line":101,"end_character":28},"updated":"2023-07-20 18:05:54.000000000","message":"can you bump this to 3, like the existing multi-node job?\n\nI wonder if we can take the higher concurrency now since we\u0027re spreading resources across two compute nodes..","commit_id":"7a183cd726ce3b46b30234d47790bf993e0822bf"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"d3d4316049786147a989ab8429d153c119460f6b","unresolved":false,"context_lines":[{"line_number":98,"context_line":"        CEPHADM_DEPLOY: True"},{"line_number":99,"context_line":"        MYSQL_REDUCE_MEMORY: True"},{"line_number":100,"context_line":"        REMOTE_CEPH: False"},{"line_number":101,"context_line":"      tempest_concurrency: 1"},{"line_number":102,"context_line":"    group-vars:"},{"line_number":103,"context_line":"      subnode:"},{"line_number":104,"context_line":"        devstack_localrc:"}],"source_content_type":"text/x-yaml","patch_set":41,"id":"1b2040c6_3e6d7dea","line":101,"range":{"start_line":101,"start_character":6,"end_line":101,"end_character":28},"in_reply_to":"7d260b5e_f000fc73","updated":"2023-07-20 23:54:29.000000000","message":"Done","commit_id":"7a183cd726ce3b46b30234d47790bf993e0822bf"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"d3d4316049786147a989ab8429d153c119460f6b","unresolved":true,"context_lines":[{"line_number":37,"context_line":"      tempest_test_blacklist: \u0027{{ ansible_user_dir }}/{{ zuul.projects[\"opendev.org/openstack/devstack-plugin-ceph\"].src_dir }}/tempest_skiplist.txt\u0027"},{"line_number":38,"context_line":""},{"line_number":39,"context_line":"- job:"},{"line_number":40,"context_line":"    name: devstack-plugin-ceph-tempest-ubuntu"},{"line_number":41,"context_line":"    parent: devstack-plugin-ceph-tempest-py3-base"},{"line_number":42,"context_line":"    description: |"},{"line_number":43,"context_line":"      Integration tests that runs with the ceph devstack plugin using ceph"}],"source_content_type":"text/x-yaml","patch_set":43,"id":"7677dc00_3e40e1ef","side":"PARENT","line":40,"updated":"2023-07-20 23:54:29.000000000","message":"don\u0027t remove these jobs; lets do any such cleanup in a different change","commit_id":"bdce6ce508671b9fb6be7825c8c244e742895362"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"68306ddff5589a15ab52858425967720a3083d15","unresolved":false,"context_lines":[{"line_number":37,"context_line":"      tempest_test_blacklist: \u0027{{ ansible_user_dir }}/{{ zuul.projects[\"opendev.org/openstack/devstack-plugin-ceph\"].src_dir }}/tempest_skiplist.txt\u0027"},{"line_number":38,"context_line":""},{"line_number":39,"context_line":"- job:"},{"line_number":40,"context_line":"    name: devstack-plugin-ceph-tempest-ubuntu"},{"line_number":41,"context_line":"    parent: devstack-plugin-ceph-tempest-py3-base"},{"line_number":42,"context_line":"    description: |"},{"line_number":43,"context_line":"      Integration tests that runs with the ceph devstack plugin using ceph"}],"source_content_type":"text/x-yaml","patch_set":43,"id":"53e96753_1253fbce","side":"PARENT","line":40,"in_reply_to":"7677dc00_3e40e1ef","updated":"2023-07-24 19:07:16.000000000","message":"Done","commit_id":"bdce6ce508671b9fb6be7825c8c244e742895362"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"d3d4316049786147a989ab8429d153c119460f6b","unresolved":true,"context_lines":[{"line_number":146,"context_line":"          CINDER_CEPH_UUID: d531d2d4-3937-429c-b0c2-658fe41e82aa"},{"line_number":147,"context_line":""},{"line_number":148,"context_line":"- job:"},{"line_number":149,"context_line":"    name: devstack-plugin-ceph-master-tempest"},{"line_number":150,"context_line":"    parent: devstack-plugin-ceph-tempest-py3-base"},{"line_number":151,"context_line":"    description: |"},{"line_number":152,"context_line":"      Integration tests that runs with the ceph devstack plugin"}],"source_content_type":"text/x-yaml","patch_set":43,"id":"392f5eb4_f5ae165f","side":"PARENT","line":149,"updated":"2023-07-20 23:54:29.000000000","message":"same comment as above..","commit_id":"bdce6ce508671b9fb6be7825c8c244e742895362"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"68306ddff5589a15ab52858425967720a3083d15","unresolved":false,"context_lines":[{"line_number":146,"context_line":"          CINDER_CEPH_UUID: d531d2d4-3937-429c-b0c2-658fe41e82aa"},{"line_number":147,"context_line":""},{"line_number":148,"context_line":"- job:"},{"line_number":149,"context_line":"    name: devstack-plugin-ceph-master-tempest"},{"line_number":150,"context_line":"    parent: devstack-plugin-ceph-tempest-py3-base"},{"line_number":151,"context_line":"    description: |"},{"line_number":152,"context_line":"      Integration tests that runs with the ceph devstack plugin"}],"source_content_type":"text/x-yaml","patch_set":43,"id":"26dce27b_33af3e38","side":"PARENT","line":149,"in_reply_to":"392f5eb4_f5ae165f","updated":"2023-07-24 19:07:16.000000000","message":"Done","commit_id":"bdce6ce508671b9fb6be7825c8c244e742895362"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"d3d4316049786147a989ab8429d153c119460f6b","unresolved":true,"context_lines":[{"line_number":163,"context_line":"    check:"},{"line_number":164,"context_line":"      jobs:"},{"line_number":165,"context_line":"        - devstack-plugin-ceph-tempest-py3"},{"line_number":166,"context_line":"        - devstack-plugin-ceph-multinode-tempest-py3:"},{"line_number":167,"context_line":"            voting: false"},{"line_number":168,"context_line":"        - devstack-plugin-ceph-multinode-tempest-cephadm:"},{"line_number":169,"context_line":"            voting: false"},{"line_number":170,"context_line":"        - devstack-plugin-ceph-cephfs-nfs:"}],"source_content_type":"text/x-yaml","patch_set":43,"id":"b75122a0_8c0245e3","line":167,"range":{"start_line":166,"start_character":10,"end_line":167,"end_character":25},"updated":"2023-07-20 23:54:29.000000000","message":"can you please replace this job to use cephadm instead? you can delete the other job altogether; and this can remain \"non-voting\"\n\n\nSo the intent of this change is to replace the existing multinode job to use cephadm instead of the package based installation. That\u0027s what we\u0027ve essentially done with the single node job above (\"devstack-plugin-ceph-tempest-py3\")","commit_id":"c4486eb7b79a9a15c5b80d301ebd23681d9d4030"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"68306ddff5589a15ab52858425967720a3083d15","unresolved":false,"context_lines":[{"line_number":163,"context_line":"    check:"},{"line_number":164,"context_line":"      jobs:"},{"line_number":165,"context_line":"        - devstack-plugin-ceph-tempest-py3"},{"line_number":166,"context_line":"        - devstack-plugin-ceph-multinode-tempest-py3:"},{"line_number":167,"context_line":"            voting: false"},{"line_number":168,"context_line":"        - devstack-plugin-ceph-multinode-tempest-cephadm:"},{"line_number":169,"context_line":"            voting: false"},{"line_number":170,"context_line":"        - devstack-plugin-ceph-cephfs-nfs:"}],"source_content_type":"text/x-yaml","patch_set":43,"id":"659a272a_e58a618b","line":167,"range":{"start_line":166,"start_character":10,"end_line":167,"end_character":25},"in_reply_to":"b75122a0_8c0245e3","updated":"2023-07-24 19:07:16.000000000","message":"Done","commit_id":"c4486eb7b79a9a15c5b80d301ebd23681d9d4030"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"9bb229063f860b85365ff0b8e375e0ea2cbaf055","unresolved":true,"context_lines":[{"line_number":168,"context_line":"    check:"},{"line_number":169,"context_line":"      jobs:"},{"line_number":170,"context_line":"        - devstack-plugin-ceph-tempest-py3"},{"line_number":171,"context_line":"        - devstack-plugin-ceph-tempest-ubuntu"},{"line_number":172,"context_line":"        - devstack-plugin-ceph-multinode-tempest-py3:"},{"line_number":173,"context_line":"            voting: false"},{"line_number":174,"context_line":"        - devstack-plugin-ceph-cephfs-native:"}],"source_content_type":"text/x-yaml","patch_set":44,"id":"41521199_64637bfd","line":171,"range":{"start_line":171,"start_character":0,"end_line":171,"end_character":45},"updated":"2023-07-24 20:17:56.000000000","message":"this job is nonvoting currently; please keep it so","commit_id":"96ef1e49d06839928372b1b99f8d0b60bbfa8df4"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b681f164ec6123dc86767b7ca20734823d01dc22","unresolved":false,"context_lines":[{"line_number":168,"context_line":"    check:"},{"line_number":169,"context_line":"      jobs:"},{"line_number":170,"context_line":"        - devstack-plugin-ceph-tempest-py3"},{"line_number":171,"context_line":"        - devstack-plugin-ceph-tempest-ubuntu"},{"line_number":172,"context_line":"        - devstack-plugin-ceph-multinode-tempest-py3:"},{"line_number":173,"context_line":"            voting: false"},{"line_number":174,"context_line":"        - devstack-plugin-ceph-cephfs-native:"}],"source_content_type":"text/x-yaml","patch_set":44,"id":"6ffd57ae_1fd58697","line":171,"range":{"start_line":171,"start_character":0,"end_line":171,"end_character":45},"in_reply_to":"41521199_64637bfd","updated":"2023-08-07 20:11:14.000000000","message":"Done","commit_id":"96ef1e49d06839928372b1b99f8d0b60bbfa8df4"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"9bb229063f860b85365ff0b8e375e0ea2cbaf055","unresolved":true,"context_lines":[{"line_number":171,"context_line":"        - devstack-plugin-ceph-tempest-ubuntu"},{"line_number":172,"context_line":"        - devstack-plugin-ceph-multinode-tempest-py3:"},{"line_number":173,"context_line":"            voting: false"},{"line_number":174,"context_line":"        - devstack-plugin-ceph-cephfs-native:"},{"line_number":175,"context_line":"            irrelevant-files: *irrelevant-files"},{"line_number":176,"context_line":"        - devstack-plugin-ceph-cephfs-nfs:"},{"line_number":177,"context_line":"            irrelevant-files: *irrelevant-files"},{"line_number":178,"context_line":"            voting: false"}],"source_content_type":"text/x-yaml","patch_set":44,"id":"fc8a02de_f92daebe","line":175,"range":{"start_line":174,"start_character":0,"end_line":175,"end_character":47},"updated":"2023-07-24 20:17:56.000000000","message":"this job is nonvoting currently; please keep it so","commit_id":"96ef1e49d06839928372b1b99f8d0b60bbfa8df4"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b681f164ec6123dc86767b7ca20734823d01dc22","unresolved":false,"context_lines":[{"line_number":171,"context_line":"        - devstack-plugin-ceph-tempest-ubuntu"},{"line_number":172,"context_line":"        - devstack-plugin-ceph-multinode-tempest-py3:"},{"line_number":173,"context_line":"            voting: false"},{"line_number":174,"context_line":"        - devstack-plugin-ceph-cephfs-native:"},{"line_number":175,"context_line":"            irrelevant-files: *irrelevant-files"},{"line_number":176,"context_line":"        - devstack-plugin-ceph-cephfs-nfs:"},{"line_number":177,"context_line":"            irrelevant-files: *irrelevant-files"},{"line_number":178,"context_line":"            voting: false"}],"source_content_type":"text/x-yaml","patch_set":44,"id":"62d7e6aa_242f6955","line":175,"range":{"start_line":174,"start_character":0,"end_line":175,"end_character":47},"in_reply_to":"fc8a02de_f92daebe","updated":"2023-08-07 20:11:14.000000000","message":"Done","commit_id":"96ef1e49d06839928372b1b99f8d0b60bbfa8df4"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"24dae116705ed203d0279ca17e370925b3d4d8b0","unresolved":true,"context_lines":[{"line_number":92,"context_line":"    description: |"},{"line_number":93,"context_line":"      Runs manila tempest plugin tests with CephFS via NFS-Ganesha as a manila"},{"line_number":94,"context_line":"      back end (DHSS\u003dFalse)"},{"line_number":95,"context_line":"    parent: manila-tempest-plugin-cephfs-nfs"},{"line_number":96,"context_line":"    nodeset: devstack-single-node-centos-9-stream"},{"line_number":97,"context_line":"    vars:"},{"line_number":98,"context_line":"      # TODO(gouthamr): some tests are disabled due to bugs"},{"line_number":99,"context_line":"      # IPv6 Tests: https://bugs.launchpad.net/manila/+bug/1998489"},{"line_number":100,"context_line":"      # snapshot clone fs sync: https://bugs.launchpad.net/manila/+bug/1989273"},{"line_number":101,"context_line":"      tempest_exclude_regex: \"\\"},{"line_number":102,"context_line":"      (^manila_tempest_tests.tests.scenario.*IPv6.*)|\\"},{"line_number":103,"context_line":"      (^manila_tempest_tests.tests.scenario.test_share_basic_ops.TestShareBasicOpsNFS.test_write_data_to_share_created_from_snapshot)\""},{"line_number":104,"context_line":"      devstack_localrc:"},{"line_number":105,"context_line":"        MANILA_OPTGROUP_cephfsnfs_cephfs_ganesha_server_ip: \"{{ hostvars[inventory_hostname][\u0027nodepool\u0027][\u0027private_ipv4\u0027] }}\""},{"line_number":106,"context_line":"        CEPH_RELEASE: \"quincy\""},{"line_number":107,"context_line":"        MANILA_SETUP_IPV6: false"},{"line_number":108,"context_line":"        NEUTRON_CREATE_INITIAL_NETWORKS: true"},{"line_number":109,"context_line":"        IP_VERSION: 4"},{"line_number":110,"context_line":""},{"line_number":111,"context_line":"- job:"},{"line_number":112,"context_line":"    name: devstack-plugin-ceph-multinode-tempest-py3"}],"source_content_type":"text/x-yaml","patch_set":49,"id":"610b8d93_b0360d14","line":109,"range":{"start_line":95,"start_character":44,"end_line":109,"end_character":21},"updated":"2023-08-08 21:56:44.000000000","message":"this looks like a bad rebase; could you please undo? \n\na patch merged recently that removed these overrides: https://review.opendev.org/c/openstack/devstack-plugin-ceph/+/882625","commit_id":"4fa036dd59a0aa94e0d0d3706b8b9c168666e89a"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"23907bf81cea11ea40be4f55cb04d3a21a1db2b1","unresolved":false,"context_lines":[{"line_number":92,"context_line":"    description: |"},{"line_number":93,"context_line":"      Runs manila tempest plugin tests with CephFS via NFS-Ganesha as a manila"},{"line_number":94,"context_line":"      back end (DHSS\u003dFalse)"},{"line_number":95,"context_line":"    parent: manila-tempest-plugin-cephfs-nfs"},{"line_number":96,"context_line":"    nodeset: devstack-single-node-centos-9-stream"},{"line_number":97,"context_line":"    vars:"},{"line_number":98,"context_line":"      # TODO(gouthamr): some tests are disabled due to bugs"},{"line_number":99,"context_line":"      # IPv6 Tests: https://bugs.launchpad.net/manila/+bug/1998489"},{"line_number":100,"context_line":"      # snapshot clone fs sync: https://bugs.launchpad.net/manila/+bug/1989273"},{"line_number":101,"context_line":"      tempest_exclude_regex: \"\\"},{"line_number":102,"context_line":"      (^manila_tempest_tests.tests.scenario.*IPv6.*)|\\"},{"line_number":103,"context_line":"      (^manila_tempest_tests.tests.scenario.test_share_basic_ops.TestShareBasicOpsNFS.test_write_data_to_share_created_from_snapshot)\""},{"line_number":104,"context_line":"      devstack_localrc:"},{"line_number":105,"context_line":"        MANILA_OPTGROUP_cephfsnfs_cephfs_ganesha_server_ip: \"{{ hostvars[inventory_hostname][\u0027nodepool\u0027][\u0027private_ipv4\u0027] }}\""},{"line_number":106,"context_line":"        CEPH_RELEASE: \"quincy\""},{"line_number":107,"context_line":"        MANILA_SETUP_IPV6: false"},{"line_number":108,"context_line":"        NEUTRON_CREATE_INITIAL_NETWORKS: true"},{"line_number":109,"context_line":"        IP_VERSION: 4"},{"line_number":110,"context_line":""},{"line_number":111,"context_line":"- job:"},{"line_number":112,"context_line":"    name: devstack-plugin-ceph-multinode-tempest-py3"}],"source_content_type":"text/x-yaml","patch_set":49,"id":"430f617c_89a5f93d","line":109,"range":{"start_line":95,"start_character":44,"end_line":109,"end_character":21},"in_reply_to":"610b8d93_b0360d14","updated":"2023-08-11 15:34:08.000000000","message":"Done","commit_id":"4fa036dd59a0aa94e0d0d3706b8b9c168666e89a"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"05f5e39baa8cf342ab3ca59bd95b57770896eeab","unresolved":true,"context_lines":[{"line_number":157,"context_line":"        - devstack-plugin-ceph-tempest-ubuntu:"},{"line_number":158,"context_line":"            voting: false"},{"line_number":159,"context_line":"        - devstack-plugin-ceph-multinode-tempest-py3:"},{"line_number":160,"context_line":"            voting: false"},{"line_number":161,"context_line":"        - devstack-plugin-ceph-cephfs-native:"},{"line_number":162,"context_line":"            irrelevant-files: *irrelevant-files"},{"line_number":163,"context_line":"            voting: false"}],"source_content_type":"text/x-yaml","patch_set":50,"id":"dc2dbfae_3fabcb9d","line":160,"updated":"2023-08-22 21:44:00.000000000","message":"This should be voting, no? We have jobs that depend on this that were voting in the past and got broken because this isn\u0027t voting for the repo in which it is defined. We\u0027ve since make it nonvoting while we wait, but we\u0027d definitely like to see this go back.\n\nIs there something we\u0027re waiting for? If it\u0027s soak time, perhaps we could queue up a patch behind this to make it voting that we can merge soon? And if not, let\u0027s make it voting now :)","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"7c5892e8b960f58e1ea9c676ff4ecd05f13fd566","unresolved":false,"context_lines":[{"line_number":157,"context_line":"        - devstack-plugin-ceph-tempest-ubuntu:"},{"line_number":158,"context_line":"            voting: false"},{"line_number":159,"context_line":"        - devstack-plugin-ceph-multinode-tempest-py3:"},{"line_number":160,"context_line":"            voting: false"},{"line_number":161,"context_line":"        - devstack-plugin-ceph-cephfs-native:"},{"line_number":162,"context_line":"            irrelevant-files: *irrelevant-files"},{"line_number":163,"context_line":"            voting: false"}],"source_content_type":"text/x-yaml","patch_set":50,"id":"295ebc3a_4e8032c3","line":160,"in_reply_to":"b55dda40_77c2e702","updated":"2023-09-07 20:16:35.000000000","message":"Ashley and I spoke about this; we\u0027re quite close to RC1.. so we agreed to convert this to voting right away..","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"47a9f72fbb6ab237d1ee0db2d9e9d10c78365fc6","unresolved":true,"context_lines":[{"line_number":157,"context_line":"        - devstack-plugin-ceph-tempest-ubuntu:"},{"line_number":158,"context_line":"            voting: false"},{"line_number":159,"context_line":"        - devstack-plugin-ceph-multinode-tempest-py3:"},{"line_number":160,"context_line":"            voting: false"},{"line_number":161,"context_line":"        - devstack-plugin-ceph-cephfs-native:"},{"line_number":162,"context_line":"            irrelevant-files: *irrelevant-files"},{"line_number":163,"context_line":"            voting: false"}],"source_content_type":"text/x-yaml","patch_set":50,"id":"b55dda40_77c2e702","line":160,"in_reply_to":"d1f4c1bd_7adcc795","updated":"2023-08-22 22:14:22.000000000","message":"Yeah, we had a job based on this in nova which was voting, and it took us a while to figure out how we could be broken and the plugin repo was not... 😊\n\n\nCool, yep, if you\u0027ll do that and we can merge it around rc time, that sounds good to me, thanks!","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"a2022dee88f9a6dce1de8b089b9e8807f642ba68","unresolved":true,"context_lines":[{"line_number":157,"context_line":"        - devstack-plugin-ceph-tempest-ubuntu:"},{"line_number":158,"context_line":"            voting: false"},{"line_number":159,"context_line":"        - devstack-plugin-ceph-multinode-tempest-py3:"},{"line_number":160,"context_line":"            voting: false"},{"line_number":161,"context_line":"        - devstack-plugin-ceph-cephfs-native:"},{"line_number":162,"context_line":"            irrelevant-files: *irrelevant-files"},{"line_number":163,"context_line":"            voting: false"}],"source_content_type":"text/x-yaml","patch_set":50,"id":"d1f4c1bd_7adcc795","line":160,"in_reply_to":"dc2dbfae_3fabcb9d","updated":"2023-08-22 22:02:59.000000000","message":"Ack; thanks for sharing the history on this. I agree, if voting jobs depend on this, we should keep this voting. My objective with asking for it to be \"non-voting\" is to match what we had in this repo - but i do realize we probably switched it to non-voting to buy time to fix it in the first place :) \n\nI think we give this a bit of soaktime and then switch to voting. We could do this around Bobcat rc; wdyt?\n\nIn that time; i\u0027ll do a deprecation patch for the package based deploy that will flip this deployment mode on by default (\"CEPHADM_DEPLOY\" will default to True) and write an email to the ML.","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"7c5892e8b960f58e1ea9c676ff4ecd05f13fd566","unresolved":true,"context_lines":[{"line_number":93,"context_line":"      Runs manila tempest plugin tests with CephFS via NFS-Ganesha as a manila"},{"line_number":94,"context_line":"      back end (DHSS\u003dFalse)"},{"line_number":95,"context_line":"    parent: manila-tempest-plugin-cephfs-nfs"},{"line_number":96,"context_line":"    nodeset: devstack-single-node-centos-9-stream"},{"line_number":97,"context_line":"    vars:"},{"line_number":98,"context_line":"      # TODO(gouthamr): some tests are disabled due to bugs"},{"line_number":99,"context_line":"      # IPv6 Tests: https://bugs.launchpad.net/manila/+bug/1998489"},{"line_number":100,"context_line":"      # snapshot clone fs sync: https://bugs.launchpad.net/manila/+bug/1989273"},{"line_number":101,"context_line":"      tempest_exclude_regex: \"\\"},{"line_number":102,"context_line":"      (^manila_tempest_tests.tests.scenario.*IPv6.*)|\\"},{"line_number":103,"context_line":"      (^manila_tempest_tests.tests.scenario.test_share_basic_ops.TestShareBasicOpsNFS.test_write_data_to_share_created_from_snapshot)\""},{"line_number":104,"context_line":"      devstack_localrc:"},{"line_number":105,"context_line":"        MANILA_OPTGROUP_cephfsnfs_cephfs_ganesha_server_ip: \"{{ hostvars[inventory_hostname][\u0027nodepool\u0027][\u0027private_ipv4\u0027] }}\""},{"line_number":106,"context_line":"        CEPH_RELEASE: \"quincy\""},{"line_number":107,"context_line":"        MANILA_SETUP_IPV6: false"},{"line_number":108,"context_line":"        NEUTRON_CREATE_INITIAL_NETWORKS: true"},{"line_number":109,"context_line":"        IP_VERSION: 4"},{"line_number":110,"context_line":""},{"line_number":111,"context_line":"- job:"},{"line_number":112,"context_line":"    name: devstack-plugin-ceph-multinode-tempest-py3"}],"source_content_type":"text/x-yaml","patch_set":52,"id":"9884030e_39d825c8","line":109,"range":{"start_line":96,"start_character":0,"end_line":109,"end_character":21},"updated":"2023-09-07 20:16:35.000000000","message":"this stuff came back; bad rebase i guess.. please drop..","commit_id":"024e2d2c1188c3e058abf130995e858129449cc8"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"fee6eca750c0820f60e70bfc62cbd5e1a236170b","unresolved":false,"context_lines":[{"line_number":93,"context_line":"      Runs manila tempest plugin tests with CephFS via NFS-Ganesha as a manila"},{"line_number":94,"context_line":"      back end (DHSS\u003dFalse)"},{"line_number":95,"context_line":"    parent: manila-tempest-plugin-cephfs-nfs"},{"line_number":96,"context_line":"    nodeset: devstack-single-node-centos-9-stream"},{"line_number":97,"context_line":"    vars:"},{"line_number":98,"context_line":"      # TODO(gouthamr): some tests are disabled due to bugs"},{"line_number":99,"context_line":"      # IPv6 Tests: https://bugs.launchpad.net/manila/+bug/1998489"},{"line_number":100,"context_line":"      # snapshot clone fs sync: https://bugs.launchpad.net/manila/+bug/1989273"},{"line_number":101,"context_line":"      tempest_exclude_regex: \"\\"},{"line_number":102,"context_line":"      (^manila_tempest_tests.tests.scenario.*IPv6.*)|\\"},{"line_number":103,"context_line":"      (^manila_tempest_tests.tests.scenario.test_share_basic_ops.TestShareBasicOpsNFS.test_write_data_to_share_created_from_snapshot)\""},{"line_number":104,"context_line":"      devstack_localrc:"},{"line_number":105,"context_line":"        MANILA_OPTGROUP_cephfsnfs_cephfs_ganesha_server_ip: \"{{ hostvars[inventory_hostname][\u0027nodepool\u0027][\u0027private_ipv4\u0027] }}\""},{"line_number":106,"context_line":"        CEPH_RELEASE: \"quincy\""},{"line_number":107,"context_line":"        MANILA_SETUP_IPV6: false"},{"line_number":108,"context_line":"        NEUTRON_CREATE_INITIAL_NETWORKS: true"},{"line_number":109,"context_line":"        IP_VERSION: 4"},{"line_number":110,"context_line":""},{"line_number":111,"context_line":"- job:"},{"line_number":112,"context_line":"    name: devstack-plugin-ceph-multinode-tempest-py3"}],"source_content_type":"text/x-yaml","patch_set":52,"id":"e5bac7a5_08e3dad0","line":109,"range":{"start_line":96,"start_character":0,"end_line":109,"end_character":21},"in_reply_to":"9884030e_39d825c8","updated":"2023-09-12 20:52:48.000000000","message":"Done","commit_id":"024e2d2c1188c3e058abf130995e858129449cc8"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"f69329dd889f09dc9d661faac47e5aed43ce6e91","unresolved":true,"context_lines":[{"line_number":104,"context_line":"      - openstack/cinder-tempest-plugin"},{"line_number":105,"context_line":"      - openstack/devstack-plugin-ceph"},{"line_number":106,"context_line":"    timeout: 10800"},{"line_number":107,"context_line":"    voting: false"},{"line_number":108,"context_line":"    vars:"},{"line_number":109,"context_line":"      configure_swap_size: 8192"},{"line_number":110,"context_line":"      tempest_concurrency: 3"}],"source_content_type":"text/x-yaml","patch_set":53,"id":"1cb826b2_f4e75ae5","line":107,"updated":"2023-09-11 00:50:49.000000000","message":"can you please remove this?","commit_id":"7c27f63e42e5e8fde6ff54b757c46c43674c5c51"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"fee6eca750c0820f60e70bfc62cbd5e1a236170b","unresolved":false,"context_lines":[{"line_number":104,"context_line":"      - openstack/cinder-tempest-plugin"},{"line_number":105,"context_line":"      - openstack/devstack-plugin-ceph"},{"line_number":106,"context_line":"    timeout: 10800"},{"line_number":107,"context_line":"    voting: false"},{"line_number":108,"context_line":"    vars:"},{"line_number":109,"context_line":"      configure_swap_size: 8192"},{"line_number":110,"context_line":"      tempest_concurrency: 3"}],"source_content_type":"text/x-yaml","patch_set":53,"id":"1fe66d3f_850d3c78","line":107,"in_reply_to":"1cb826b2_f4e75ae5","updated":"2023-09-12 20:52:48.000000000","message":"Done","commit_id":"7c27f63e42e5e8fde6ff54b757c46c43674c5c51"},{"author":{"_account_id":19262,"name":"Liron Kuchlani","email":"lkuchlan@redhat.com","username":"lkuchlan"},"change_message_id":"0391738065255d4979924c271f1af8420b59cd81","unresolved":true,"context_lines":[{"line_number":107,"context_line":"    voting: false"},{"line_number":108,"context_line":"    vars:"},{"line_number":109,"context_line":"      configure_swap_size: 8192"},{"line_number":110,"context_line":"      tempest_concurrency: 3"},{"line_number":111,"context_line":"      devstack_localrc:"},{"line_number":112,"context_line":"        ENABLE_FILE_INJECTION: false"},{"line_number":113,"context_line":"        ENABLE_VOLUME_MULTIATTACH: true"}],"source_content_type":"text/x-yaml","patch_set":53,"id":"196ab989_6e92de74","line":110,"range":{"start_line":110,"start_character":6,"end_line":110,"end_character":28},"updated":"2023-09-11 07:34:20.000000000","message":"There\u0027s a specific rationale behind the choice of 3 instead of the default value, which typically corresponds to the number of CPUs ?","commit_id":"7c27f63e42e5e8fde6ff54b757c46c43674c5c51"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"fee6eca750c0820f60e70bfc62cbd5e1a236170b","unresolved":true,"context_lines":[{"line_number":107,"context_line":"    voting: false"},{"line_number":108,"context_line":"    vars:"},{"line_number":109,"context_line":"      configure_swap_size: 8192"},{"line_number":110,"context_line":"      tempest_concurrency: 3"},{"line_number":111,"context_line":"      devstack_localrc:"},{"line_number":112,"context_line":"        ENABLE_FILE_INJECTION: false"},{"line_number":113,"context_line":"        ENABLE_VOLUME_MULTIATTACH: true"}],"source_content_type":"text/x-yaml","patch_set":53,"id":"884d062b_6b295579","line":110,"range":{"start_line":110,"start_character":6,"end_line":110,"end_character":28},"in_reply_to":"196ab989_6e92de74","updated":"2023-09-12 20:52:48.000000000","message":"I think I picked this number up from some previous iteration of this job, theres no particular rationale beihnd this.","commit_id":"7c27f63e42e5e8fde6ff54b757c46c43674c5c51"}],"/COMMIT_MSG":[{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"5515c1bbf4800c10ca804b5d7b6411f52a069942","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":30,"id":"c2cb96a8_787213d5","line":15,"updated":"2023-06-07 19:28:36.000000000","message":"Needed-By: I19558a7e3122ef3439c39c1e41661bdae1149944","commit_id":"1c192e89fcc169e8731f80b281780b7dac8a0aaa"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"d3d4316049786147a989ab8429d153c119460f6b","unresolved":true,"context_lines":[],"source_content_type":"","patch_set":30,"id":"ea558cc4_1230b01d","line":15,"in_reply_to":"c2cb96a8_787213d5","updated":"2023-07-20 23:54:29.000000000","message":"please add this on the next patchset; helps indentify where we\u0027re going with this","commit_id":"1c192e89fcc169e8731f80b281780b7dac8a0aaa"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"68306ddff5589a15ab52858425967720a3083d15","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":30,"id":"534fa381_df93b5e1","line":15,"in_reply_to":"ea558cc4_1230b01d","updated":"2023-07-24 19:07:16.000000000","message":"Done","commit_id":"1c192e89fcc169e8731f80b281780b7dac8a0aaa"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"d3d4316049786147a989ab8429d153c119460f6b","unresolved":true,"context_lines":[{"line_number":4,"context_line":"Commit:     ashrod98 \u003cashrod98@gmail.com\u003e"},{"line_number":5,"context_line":"CommitDate: 2023-07-20 20:59:00 +0000"},{"line_number":6,"context_line":""},{"line_number":7,"context_line":"[WIP][DNM] Remote Ceph with cephadm"},{"line_number":8,"context_line":""},{"line_number":9,"context_line":"Add podman and jq as part of preinstall dependency."},{"line_number":10,"context_line":"Add REMOTE_CEPH capabilities to CEPHADM deployment."}],"source_content_type":"text/x-gerrit-commit-message","patch_set":43,"id":"6f3839db_b7aadd11","line":7,"range":{"start_line":7,"start_character":0,"end_line":7,"end_character":11},"updated":"2023-07-20 23:54:29.000000000","message":"drop","commit_id":"c4486eb7b79a9a15c5b80d301ebd23681d9d4030"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"68306ddff5589a15ab52858425967720a3083d15","unresolved":false,"context_lines":[{"line_number":4,"context_line":"Commit:     ashrod98 \u003cashrod98@gmail.com\u003e"},{"line_number":5,"context_line":"CommitDate: 2023-07-20 20:59:00 +0000"},{"line_number":6,"context_line":""},{"line_number":7,"context_line":"[WIP][DNM] Remote Ceph with cephadm"},{"line_number":8,"context_line":""},{"line_number":9,"context_line":"Add podman and jq as part of preinstall dependency."},{"line_number":10,"context_line":"Add REMOTE_CEPH capabilities to CEPHADM deployment."}],"source_content_type":"text/x-gerrit-commit-message","patch_set":43,"id":"4add9670_19cd8d6b","line":7,"range":{"start_line":7,"start_character":0,"end_line":7,"end_character":11},"in_reply_to":"6f3839db_b7aadd11","updated":"2023-07-24 19:07:16.000000000","message":"Done","commit_id":"c4486eb7b79a9a15c5b80d301ebd23681d9d4030"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"06425f9e7ff4dfd0d3be2dc6df02a1cfdafc8334","unresolved":true,"context_lines":[{"line_number":14,"context_line":""},{"line_number":15,"context_line":"Part of an effort to test multinode deployments with cephadm."},{"line_number":16,"context_line":""},{"line_number":17,"context_line":"Needed-By: I19558a7e3122ef3439c39c1e41661bdae1149944"},{"line_number":18,"context_line":"Change-Id: I84249ae268dfe00a112c67e5170b679acb318a25"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":46,"id":"9ebdc910_e9344f4d","line":17,"range":{"start_line":17,"start_character":10,"end_line":17,"end_character":52},"updated":"2023-07-25 23:40:43.000000000","message":"This should be I5162815b66d3f3e8cf8c1e246b61b0ea06c1a270 now","commit_id":"442ff1d7b0ce6347a45cbac247b34213557b710d"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b681f164ec6123dc86767b7ca20734823d01dc22","unresolved":false,"context_lines":[{"line_number":14,"context_line":""},{"line_number":15,"context_line":"Part of an effort to test multinode deployments with cephadm."},{"line_number":16,"context_line":""},{"line_number":17,"context_line":"Needed-By: I19558a7e3122ef3439c39c1e41661bdae1149944"},{"line_number":18,"context_line":"Change-Id: I84249ae268dfe00a112c67e5170b679acb318a25"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":46,"id":"3041ecc7_9ce8eb63","line":17,"range":{"start_line":17,"start_character":10,"end_line":17,"end_character":52},"in_reply_to":"9ebdc910_e9344f4d","updated":"2023-08-07 20:11:14.000000000","message":"Done","commit_id":"442ff1d7b0ce6347a45cbac247b34213557b710d"}],"/PATCHSET_LEVEL":[{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"0c0c105ee808c95e4ce8f4c50134ffd59b218daa","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":3,"id":"7e25152d_822180cb","updated":"2023-04-26 21:39:25.000000000","message":"Thanks Dan for your feedback!","commit_id":"96e0984a43ee8096b1a3cf26aafc67a13bbf55e6"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"2689704e7468f3422bb062913fcf7f872b414124","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":11,"id":"438ffe67_16a2775e","updated":"2023-05-08 17:49:26.000000000","message":"Don\u0027t you need these things in here to really know if this is working?\n\nhttps://review.opendev.org/c/openstack/devstack-plugin-ceph/+/882483/1/.zuul.yaml\n\nYou can just pull those in here, if you want, and I can point my nova test patch against this change and abandon the other.","commit_id":"07477954cecd85932112dc8dc45f2e05279b5c45"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"29812d503b48d27b14b46b07de0c645d0411c735","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":11,"id":"5037b8aa_13de8c26","in_reply_to":"438ffe67_16a2775e","updated":"2023-05-08 20:12:26.000000000","message":"I added those options to this patch, feel free to make use of it. Thanks for the heads up.","commit_id":"07477954cecd85932112dc8dc45f2e05279b5c45"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"39632bb79a1f42a981fc231a75cc8f546fb773eb","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":19,"id":"4836ec43_e7e8ba33","updated":"2023-05-19 00:52:55.000000000","message":"Hi Ashley, please see comments inline. \n\nWhile you\u0027re experimenting, could you also disable jobs here that don\u0027t test your changes.. keep the cephadm jobs so they can tell you if you\u0027re breaking anything... grep the zuul config for CEPHADM_DEPLOY; you can re-enable them once you\u0027re past the experimentation..","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"d3d4316049786147a989ab8429d153c119460f6b","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":43,"id":"e944cdab_28d433fa","updated":"2023-07-20 23:54:29.000000000","message":"Great to see this job pass! Good work on this Ashley; imo, just a bit more cleanup to close this out","commit_id":"c4486eb7b79a9a15c5b80d301ebd23681d9d4030"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"68306ddff5589a15ab52858425967720a3083d15","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":44,"id":"13a18ab7_509fe706","updated":"2023-07-24 19:07:16.000000000","message":"recheck","commit_id":"96ef1e49d06839928372b1b99f8d0b60bbfa8df4"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"9bb229063f860b85365ff0b8e375e0ea2cbaf055","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":44,"id":"19c712dc_f0bcbd32","updated":"2023-07-24 20:17:56.000000000","message":"the failures on this job currently are unrelated; could you please fix up the .zuul.conf, the jobs will be re-attempted on your next patch","commit_id":"96ef1e49d06839928372b1b99f8d0b60bbfa8df4"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"a3a1f65bf91bc1faef5f7f59f849b657ed98359f","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":46,"id":"cf6f0326_ae82001d","updated":"2023-07-24 21:52:14.000000000","message":"\u003e mapping values are not allowed in this context\n\u003e in \"openstack/devstack-plugin-ceph/.zuul.yaml@master\", line 172, column 19\n\nMade a minor edit to fix this up","commit_id":"442ff1d7b0ce6347a45cbac247b34213557b710d"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"9d9509dcabaa7851e3b5c2b6511f47c4709e8a7e","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":46,"id":"79c4772a_ef760f96","updated":"2023-07-26 19:45:33.000000000","message":"I haven\u0027t looked at this in detail, but here are a few comments from a skim. Probably need to get some of the devstack heavy(er) hitter to look at this before a nobody like me :)","commit_id":"442ff1d7b0ce6347a45cbac247b34213557b710d"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"c34958ec37e79c2a2903106aab71fe6643bd7a4c","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":46,"id":"01cf61ae_3da6cfcd","updated":"2023-07-25 23:39:53.000000000","message":"Works; LGTM, thank you!","commit_id":"442ff1d7b0ce6347a45cbac247b34213557b710d"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b681f164ec6123dc86767b7ca20734823d01dc22","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":47,"id":"1bca2dd2_859ac0c0","updated":"2023-08-07 20:11:14.000000000","message":"Thanks for the reviews!","commit_id":"a424a5eaf0b2afd74f5db65773366793ed7c2a84"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"43c980ee72c6465317f6251047a803d9b1935537","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":50,"id":"646b6651_af7ab382","updated":"2023-08-11 19:24:04.000000000","message":"LGTM; thanks for the changes Ashley!","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"05f5e39baa8cf342ab3ca59bd95b57770896eeab","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":50,"id":"546740d5_cfead194","updated":"2023-08-22 21:44:00.000000000","message":"Looks like this is working and that the subnode is connected to ceph as expected, so that\u0027s good. I see in the history that Goutham asked for the job to be non-voting, but I\u0027d like to know the reason why. If we can make it voting I think we should, and if not, let\u0027s please queue up that change so it\u0027s on the radar and we can queue up changes in other projects that depend on it.\n\nOne question about the get_or_create change that can probably be just fixed with a comment (or maybe removed altogether). Those two points aside, I\u0027m good, thanks.","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"7da9185d6de8519bc9e22e92e1be6c2cc9203a7e","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":50,"id":"08956e27_eb150f29","updated":"2023-08-11 17:00:36.000000000","message":"recheck\n\nJob Timed out; a bunch of 404s in the logs.. lets see how the rerun fares","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"ca6ea523bf6b6607820c907b313cb60eb7b3f35b","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":53,"id":"5e9c3fc7_b0f6bb56","updated":"2023-09-09 16:20:29.000000000","message":"recheck\n\nZuul\u0027s log server seems down atm","commit_id":"7c27f63e42e5e8fde6ff54b757c46c43674c5c51"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"f69329dd889f09dc9d661faac47e5aed43ce6e91","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":53,"id":"a284297e_b92e6065","updated":"2023-09-11 00:50:49.000000000","message":"the job\u0027s still non-voting... \n\nbut i\u0027m a little worried that this is flaky too..","commit_id":"7c27f63e42e5e8fde6ff54b757c46c43674c5c51"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"2e082dc66be32bdb59187a32f26d90fb32dd6c34","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":54,"id":"8b31171c_cbc2491d","updated":"2023-09-13 18:26:40.000000000","message":"\"test_live_migration_with_trunk\" is failing because of a bug.\n\nIt appears that there\u0027s some issue with OVN while updating the port statuses for the sub ports [1][2]. Several fixes were attemped in Neutron, but a part of those fixes was rolled back [3][4]. I\u0027ve asked on https://review.opendev.org/c/openstack/tempest/+/888770 if we can mark the test unstable again while they investigate. \n\n[1] https://bugs.launchpad.net/neutron/+bug/2024160\n[2] https://bugs.launchpad.net/neutron/+bug/2027605\n[3] https://bugs.launchpad.net/neutron/+bug/2033887\n[4] https://review.opendev.org/c/openstack/neutron/+/893552","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"},{"author":{"_account_id":25402,"name":"Francesco Pantano","email":"fpantano@redhat.com","username":"fmount"},"change_message_id":"102db741b4a600a57936a104a830ad2035f4cd74","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":54,"id":"c246fb5d_762e9eb0","updated":"2023-09-15 18:37:24.000000000","message":"I just had a minor comment on the `min_client_version` function call, but it\u0027s really out of my curiosity and a nit, otherwise I\u0027m +2 here","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"def1b5c98a27957aba8dd2c9cee0e72350821c25","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":54,"id":"3f8c7712_690a05b5","updated":"2023-09-18 16:38:16.000000000","message":"If you have to respin I agree with the removal of the blank line and I think it\u0027s also good to have Francesco\u0027s question answered here, even if just for posterity.","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"},{"author":{"_account_id":29632,"name":"Carlos Eduardo","email":"ces.eduardo98@gmail.com","username":"silvacarlos"},"change_message_id":"132d667b4ea6f69d1a0667a4a46d113f0bcddb2c","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":54,"id":"bc3a08ad_398e4a00","updated":"2023-09-15 17:07:43.000000000","message":"LGTM, thanks Ashley","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"3c3d28b7cbaf694740f81285a229229f378dda73","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":54,"id":"ce8bc905_6b466cf9","updated":"2023-09-12 21:51:52.000000000","message":"Thanks; the job may fail, but i doubt its something to do with the job configuration...","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"62b46d55b2bebc6a0a0bd933404653c480369856","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":54,"id":"368b7bba_295989a4","updated":"2023-09-13 14:22:29.000000000","message":"recheck","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"71fa6328466699dcd5e826f6e2af4b5833b98d22","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":54,"id":"e59c9284_24f2752a","updated":"2023-09-14 19:55:24.000000000","message":"recheck, test_live_migration_with_trunk marked as unstable","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"}],"devstack/files/debs/devstack-plugin-ceph":[{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"70b249ec46d19ac13bdab38f038e4d35689e2322","unresolved":true,"context_lines":[{"line_number":1,"context_line":"xfsprogs"},{"line_number":2,"context_line":"podman"}],"source_content_type":"application/octet-stream","patch_set":2,"id":"61f60e31_cace7311","line":2,"updated":"2023-04-26 20:29:08.000000000","message":"I also needed `catatonit` here on jammy to make it actually work. I\u0027m guessing since this is still running jobs on focal, you haven\u0027t noticed.","commit_id":"08b0c907ac1bca1fdb70e916241e0c536c7adfd4"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"0c0c105ee808c95e4ce8f4c50134ffd59b218daa","unresolved":false,"context_lines":[{"line_number":1,"context_line":"xfsprogs"},{"line_number":2,"context_line":"podman"}],"source_content_type":"application/octet-stream","patch_set":2,"id":"9aada193_0d58303e","line":2,"in_reply_to":"61f60e31_cace7311","updated":"2023-04-26 21:39:25.000000000","message":"Done","commit_id":"08b0c907ac1bca1fdb70e916241e0c536c7adfd4"}],"devstack/lib/cephadm":[{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"70b249ec46d19ac13bdab38f038e4d35689e2322","unresolved":true,"context_lines":[{"line_number":122,"context_line":"    # FIXME(vkmc) Check required for Ubuntu 20.04 LTS (current CI node)"},{"line_number":123,"context_line":"    # Remove when our CI is pushed to the next LTS version"},{"line_number":124,"context_line":"    if ! command -v podman \u0026\u003e /dev/null; then"},{"line_number":125,"context_line":"        if [[ $os_CODENAME \u003d~ (focal|jammy) ]]; then"},{"line_number":126,"context_line":"            echo \"deb https://download.opensuse.org/repositories/devel:/kubic:/libcontainers:/stable/xUbuntu_${os_RELEASE}/ /\" \\"},{"line_number":127,"context_line":"                | sudo tee /etc/apt/sources.list.d/devel:kubic:libcontainers:stable.list"},{"line_number":128,"context_line":"            curl -L \"https://download.opensuse.org/repositories/devel:/kubic:/libcontainers:/stable/xUbuntu_${os_RELEASE}/Release.key\" \\"}],"source_content_type":"application/x-shellscript","patch_set":2,"id":"03cb9473_a5d89890","line":125,"updated":"2023-04-26 20:29:08.000000000","message":"Why are you installing this repo for jammy? In my other attempts to get us on jammy, the podman they provide works fine. I would propose you leave this only for focal, but also that we snip all the focal workarounds from this on master anyway, since it\u0027s not on the PTI for 2023.2.","commit_id":"08b0c907ac1bca1fdb70e916241e0c536c7adfd4"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"0c0c105ee808c95e4ce8f4c50134ffd59b218daa","unresolved":false,"context_lines":[{"line_number":122,"context_line":"    # FIXME(vkmc) Check required for Ubuntu 20.04 LTS (current CI node)"},{"line_number":123,"context_line":"    # Remove when our CI is pushed to the next LTS version"},{"line_number":124,"context_line":"    if ! command -v podman \u0026\u003e /dev/null; then"},{"line_number":125,"context_line":"        if [[ $os_CODENAME \u003d~ (focal|jammy) ]]; then"},{"line_number":126,"context_line":"            echo \"deb https://download.opensuse.org/repositories/devel:/kubic:/libcontainers:/stable/xUbuntu_${os_RELEASE}/ /\" \\"},{"line_number":127,"context_line":"                | sudo tee /etc/apt/sources.list.d/devel:kubic:libcontainers:stable.list"},{"line_number":128,"context_line":"            curl -L \"https://download.opensuse.org/repositories/devel:/kubic:/libcontainers:/stable/xUbuntu_${os_RELEASE}/Release.key\" \\"}],"source_content_type":"application/x-shellscript","patch_set":2,"id":"08818317_f1360460","line":125,"in_reply_to":"03cb9473_a5d89890","updated":"2023-04-26 21:39:25.000000000","message":"Done","commit_id":"08b0c907ac1bca1fdb70e916241e0c536c7adfd4"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"39632bb79a1f42a981fc231a75cc8f546fb773eb","unresolved":true,"context_lines":[{"line_number":505,"context_line":"    bootstrap_config"},{"line_number":506,"context_line":"    get_cephadm"},{"line_number":507,"context_line":"    start_ceph"},{"line_number":508,"context_line":"    cat \u003c\u003cEOF \u003e $CEPH_CONF_DIR/REMOTE_TEST.txt"},{"line_number":509,"context_line":"I am waiting for Ceph..."},{"line_number":510,"context_line":"EOF"},{"line_number":511,"context_line":"}"}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"d3bf1a1e_58d1357c","line":508,"range":{"start_line":508,"start_character":31,"end_line":508,"end_character":42},"updated":"2023-05-19 00:52:55.000000000","message":"call this \"CEPH_CLUSTER_IS_READY\"","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b85a4b484c1faab9cb54db6445a734e8935848a5","unresolved":false,"context_lines":[{"line_number":505,"context_line":"    bootstrap_config"},{"line_number":506,"context_line":"    get_cephadm"},{"line_number":507,"context_line":"    start_ceph"},{"line_number":508,"context_line":"    cat \u003c\u003cEOF \u003e $CEPH_CONF_DIR/REMOTE_TEST.txt"},{"line_number":509,"context_line":"I am waiting for Ceph..."},{"line_number":510,"context_line":"EOF"},{"line_number":511,"context_line":"}"}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"5920db7f_8a6d516a","line":508,"range":{"start_line":508,"start_character":31,"end_line":508,"end_character":42},"in_reply_to":"d3bf1a1e_58d1357c","updated":"2023-05-19 19:58:59.000000000","message":"Done","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"39632bb79a1f42a981fc231a75cc8f546fb773eb","unresolved":true,"context_lines":[{"line_number":506,"context_line":"    get_cephadm"},{"line_number":507,"context_line":"    start_ceph"},{"line_number":508,"context_line":"    cat \u003c\u003cEOF \u003e $CEPH_CONF_DIR/REMOTE_TEST.txt"},{"line_number":509,"context_line":"I am waiting for Ceph..."},{"line_number":510,"context_line":"EOF"},{"line_number":511,"context_line":"}"},{"line_number":512,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"b3dad88c_8f46a0e9","line":509,"range":{"start_line":509,"start_character":0,"end_line":509,"end_character":21},"updated":"2023-05-19 00:52:55.000000000","message":"Cluster is provisioned.","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b85a4b484c1faab9cb54db6445a734e8935848a5","unresolved":false,"context_lines":[{"line_number":506,"context_line":"    get_cephadm"},{"line_number":507,"context_line":"    start_ceph"},{"line_number":508,"context_line":"    cat \u003c\u003cEOF \u003e $CEPH_CONF_DIR/REMOTE_TEST.txt"},{"line_number":509,"context_line":"I am waiting for Ceph..."},{"line_number":510,"context_line":"EOF"},{"line_number":511,"context_line":"}"},{"line_number":512,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"277f924a_18d09f10","line":509,"range":{"start_line":509,"start_character":0,"end_line":509,"end_character":21},"in_reply_to":"b3dad88c_8f46a0e9","updated":"2023-05-19 19:58:59.000000000","message":"Done","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"39632bb79a1f42a981fc231a75cc8f546fb773eb","unresolved":true,"context_lines":[{"line_number":515,"context_line":"    until ssh stack@$CEPH_IP ls /etc/ceph/REMOTE_TEST.txt \u0026\u003e /dev/null; do"},{"line_number":516,"context_line":"        sleep 1"},{"line_number":517,"context_line":"        echo -n ."},{"line_number":518,"context_line":"        (( TIMEOUT-- ))"},{"line_number":519,"context_line":"        [[ \"$TIMEOUT\" -eq 0 ]] \u0026\u0026 exit 1"},{"line_number":520,"context_line":"    done"},{"line_number":521,"context_line":"    echo"}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"e1cad01b_cfee0741","line":518,"range":{"start_line":518,"start_character":11,"end_line":518,"end_character":18},"updated":"2023-05-19 00:52:55.000000000","message":"this variable isn\u0027t defined...","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b85a4b484c1faab9cb54db6445a734e8935848a5","unresolved":false,"context_lines":[{"line_number":515,"context_line":"    until ssh stack@$CEPH_IP ls /etc/ceph/REMOTE_TEST.txt \u0026\u003e /dev/null; do"},{"line_number":516,"context_line":"        sleep 1"},{"line_number":517,"context_line":"        echo -n ."},{"line_number":518,"context_line":"        (( TIMEOUT-- ))"},{"line_number":519,"context_line":"        [[ \"$TIMEOUT\" -eq 0 ]] \u0026\u0026 exit 1"},{"line_number":520,"context_line":"    done"},{"line_number":521,"context_line":"    echo"}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"a43df592_c0f01b67","line":518,"range":{"start_line":518,"start_character":11,"end_line":518,"end_character":18},"in_reply_to":"e1cad01b_cfee0741","updated":"2023-05-19 19:58:59.000000000","message":"Done","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"39632bb79a1f42a981fc231a75cc8f546fb773eb","unresolved":true,"context_lines":[{"line_number":658,"context_line":"    add_pools"},{"line_number":659,"context_line":"    create_keys"},{"line_number":660,"context_line":"    if [[ \"$REMOTE_CEPH\" \u003d \"True\" ]]; then"},{"line_number":661,"context_line":"        scp -r stack@$CEPH_IP:/etc/ceph /etc/ceph"},{"line_number":662,"context_line":"    fi"},{"line_number":663,"context_line":"    client_config"},{"line_number":664,"context_line":"    import_libvirt_secret_ceph"}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"9a91b466_96a1a863","line":661,"range":{"start_line":661,"start_character":8,"end_line":661,"end_character":29},"updated":"2023-05-19 00:52:55.000000000","message":"We should allow the SSH user to be configured as well.. I think in Zuul/nodepool, the \"stack\" user is created as part of bootstrap, and there\u0027s no key exchange happening. You should rely on the zuul user perhaps.. So make this configurable, and plumb that in via the job config","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b85a4b484c1faab9cb54db6445a734e8935848a5","unresolved":false,"context_lines":[{"line_number":658,"context_line":"    add_pools"},{"line_number":659,"context_line":"    create_keys"},{"line_number":660,"context_line":"    if [[ \"$REMOTE_CEPH\" \u003d \"True\" ]]; then"},{"line_number":661,"context_line":"        scp -r stack@$CEPH_IP:/etc/ceph /etc/ceph"},{"line_number":662,"context_line":"    fi"},{"line_number":663,"context_line":"    client_config"},{"line_number":664,"context_line":"    import_libvirt_secret_ceph"}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"55dad7f0_433dfd53","line":661,"range":{"start_line":661,"start_character":8,"end_line":661,"end_character":29},"in_reply_to":"9a91b466_96a1a863","updated":"2023-05-19 19:58:59.000000000","message":"Done","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"af2f43f250652213dda347b3bb20c53c0f83ce74","unresolved":true,"context_lines":[{"line_number":676,"context_line":""},{"line_number":677,"context_line":"function cleanup_ceph {"},{"line_number":678,"context_line":"    # Cleanup the service."},{"line_number":679,"context_line":"    stop_ceph"},{"line_number":680,"context_line":"    delete_osd_dev"},{"line_number":681,"context_line":"    # purge ceph config file and keys"},{"line_number":682,"context_line":"    $SUDO rm -f ${CEPH_CONF_DIR}/*"},{"line_number":683,"context_line":"    if is_ceph_enabled_for_service nova; then"}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"e85f771d_ce43bd0e","line":680,"range":{"start_line":679,"start_character":0,"end_line":680,"end_character":18},"updated":"2023-05-19 00:56:25.000000000","message":"this needs to be skipped when REMOTE_CEPH is True","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b85a4b484c1faab9cb54db6445a734e8935848a5","unresolved":false,"context_lines":[{"line_number":676,"context_line":""},{"line_number":677,"context_line":"function cleanup_ceph {"},{"line_number":678,"context_line":"    # Cleanup the service."},{"line_number":679,"context_line":"    stop_ceph"},{"line_number":680,"context_line":"    delete_osd_dev"},{"line_number":681,"context_line":"    # purge ceph config file and keys"},{"line_number":682,"context_line":"    $SUDO rm -f ${CEPH_CONF_DIR}/*"},{"line_number":683,"context_line":"    if is_ceph_enabled_for_service nova; then"}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"9e5cc5cf_86d847f7","line":680,"range":{"start_line":679,"start_character":0,"end_line":680,"end_character":18},"in_reply_to":"e85f771d_ce43bd0e","updated":"2023-05-19 19:58:59.000000000","message":"Done","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"af2f43f250652213dda347b3bb20c53c0f83ce74","unresolved":true,"context_lines":[{"line_number":678,"context_line":"    # Cleanup the service."},{"line_number":679,"context_line":"    stop_ceph"},{"line_number":680,"context_line":"    delete_osd_dev"},{"line_number":681,"context_line":"    # purge ceph config file and keys"},{"line_number":682,"context_line":"    $SUDO rm -f ${CEPH_CONF_DIR}/*"},{"line_number":683,"context_line":"    if is_ceph_enabled_for_service nova; then"},{"line_number":684,"context_line":"    _undefine_virsh_secret"}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"0e464f0c_6d57f807","line":681,"updated":"2023-05-19 00:56:25.000000000","message":"You\u0027ll also need to delete the marker file you\u0027re creating here: https://review.opendev.org/c/openstack/devstack-plugin-ceph/+/876747/19/devstack/lib/cephadm#508","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b85a4b484c1faab9cb54db6445a734e8935848a5","unresolved":true,"context_lines":[{"line_number":678,"context_line":"    # Cleanup the service."},{"line_number":679,"context_line":"    stop_ceph"},{"line_number":680,"context_line":"    delete_osd_dev"},{"line_number":681,"context_line":"    # purge ceph config file and keys"},{"line_number":682,"context_line":"    $SUDO rm -f ${CEPH_CONF_DIR}/*"},{"line_number":683,"context_line":"    if is_ceph_enabled_for_service nova; then"},{"line_number":684,"context_line":"    _undefine_virsh_secret"}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"eb919c9e_7eddf420","line":681,"in_reply_to":"0e464f0c_6d57f807","updated":"2023-05-19 19:58:59.000000000","message":"unfamiliar with bash, but if its using rm -f /* is that force removing every file within that directory? if so, it should be removed automatically since Ive created it within ceph_conf_dir","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"4a2ef26f7334267c3275001ca5689b9eeb8590e7","unresolved":false,"context_lines":[{"line_number":678,"context_line":"    # Cleanup the service."},{"line_number":679,"context_line":"    stop_ceph"},{"line_number":680,"context_line":"    delete_osd_dev"},{"line_number":681,"context_line":"    # purge ceph config file and keys"},{"line_number":682,"context_line":"    $SUDO rm -f ${CEPH_CONF_DIR}/*"},{"line_number":683,"context_line":"    if is_ceph_enabled_for_service nova; then"},{"line_number":684,"context_line":"    _undefine_virsh_secret"}],"source_content_type":"application/x-shellscript","patch_set":19,"id":"2bc88c94_ead15746","line":681,"in_reply_to":"eb919c9e_7eddf420","updated":"2023-05-19 20:55:41.000000000","message":"yeah, that should suffice; thanks!","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"a427d4c612c2aaf1b2fbd178749db59c8e34d68e","unresolved":true,"context_lines":[{"line_number":133,"context_line":"    echo \"[GET CEPHADM] cephadm is now available\""},{"line_number":134,"context_line":""},{"line_number":135,"context_line":"    if [ -z \"$CEPHADM\" ]; then"},{"line_number":136,"context_line":"        CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""},{"line_number":137,"context_line":"    fi"},{"line_number":138,"context_line":"}"},{"line_number":139,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"b976f754_1c80ca65","line":136,"updated":"2023-05-23 14:19:03.000000000","message":"Why are you doing this? I think it\u0027d be better to leave it the way it is - have `CEPHADM` be the actual path to the tool and call it with `sudo` when appropriate (even if it\u0027s always).","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"13c833c643cad90c1b0c2b979aeb186704124feb","unresolved":true,"context_lines":[{"line_number":133,"context_line":"    echo \"[GET CEPHADM] cephadm is now available\""},{"line_number":134,"context_line":""},{"line_number":135,"context_line":"    if [ -z \"$CEPHADM\" ]; then"},{"line_number":136,"context_line":"        CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""},{"line_number":137,"context_line":"    fi"},{"line_number":138,"context_line":"}"},{"line_number":139,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"5f5dddaa_89b8a89c","line":136,"in_reply_to":"1145e893_12be3511","updated":"2023-05-23 21:49:42.000000000","message":"Yeah I guess I\u0027m confused about that part too. Shouldn\u0027t the subnode just need the conf and keys? Why does it need any ceph stuff installed at all? Could be totally my ignorance, but I thought we only need that on the main node where the ceph server side runs and that copying the `/etc/ceph` stuff in the ansible role (which is already happening) should be basically still applicable? Why does deploying ceph with cephadm on the main node change what we need on the subnode for nova, manila, etc?","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"b0e542a23663f6cdf8463cba9c247a6cb0b772f4","unresolved":true,"context_lines":[{"line_number":133,"context_line":"    echo \"[GET CEPHADM] cephadm is now available\""},{"line_number":134,"context_line":""},{"line_number":135,"context_line":"    if [ -z \"$CEPHADM\" ]; then"},{"line_number":136,"context_line":"        CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""},{"line_number":137,"context_line":"    fi"},{"line_number":138,"context_line":"}"},{"line_number":139,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"6bc4cf2a_93dbc1b2","line":136,"in_reply_to":"1bdb1c66_24511584","updated":"2023-05-23 20:03:52.000000000","message":"That\u0027s a good point. \n\nwe were trying to design this with devstack\u0027s install phases.. i.e., \"install\" will cause the creation of the ceph cluster and \"post-config\" will cause creation of the ceph resources and configuration required by devstack\u0027s services.. \n\nansiblizing this would be cleaner; but if we stick to this approach, it would require us to enhance devstack so that \"stack.sh\" can be invoked to run those install phases on demand: \n\n\n  - ./stack.sh -p override_defaults\n  - ./stack.sh -p source\n  - ./stack.sh -p pre-install\n  - ./stack.sh -p install \n  - wait for ceph to be ready\n  - ./stack.sh -p post-config\n  - ./stack.sh -p extra\n  - ./stack.sh -p test-config\n \n\nThis would be a good improvement for CI, but also limit our ability to do this locally without similar smarts","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":4146,"name":"Clark Boylan","email":"cboylan@sapwetik.org","username":"cboylan"},"change_message_id":"de990173690df158b2f76cb1b231fec8eff8e178","unresolved":true,"context_lines":[{"line_number":133,"context_line":"    echo \"[GET CEPHADM] cephadm is now available\""},{"line_number":134,"context_line":""},{"line_number":135,"context_line":"    if [ -z \"$CEPHADM\" ]; then"},{"line_number":136,"context_line":"        CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""},{"line_number":137,"context_line":"    fi"},{"line_number":138,"context_line":"}"},{"line_number":139,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"1bdb1c66_24511584","line":136,"in_reply_to":"24cf2225_bc832786","updated":"2023-05-23 17:45:56.000000000","message":"When devstack-gate and devstack grew multinode support one of the first things we did was give up trying to manually ssh back and forth between nodes and instead rely on ansible for that communications. The major upside for this is you can pretty easily express things like ordering and common tasks without repeating yourself and needing variable overrides like this.\n\nI would recommend you look into using ansible for this instead.\n\nWhere it might get a bit weird is that I think ansible for the most part is driving the devstack shell scripts. What this means is you don\u0027t really write ansible in here instead you would need to write the plugin in such a way that when ansible triggers devstack shell on the controller and computes and storage nodes (I don\u0027t know the actual layout these days) these scripts need to do the correct thing. Basically remove explicit ssh from here and instead operate under the ansible orchestration that is already occuring.","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"f9d6ae915b480f7615313fd75846cce16d273312","unresolved":true,"context_lines":[{"line_number":133,"context_line":"    echo \"[GET CEPHADM] cephadm is now available\""},{"line_number":134,"context_line":""},{"line_number":135,"context_line":"    if [ -z \"$CEPHADM\" ]; then"},{"line_number":136,"context_line":"        CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""},{"line_number":137,"context_line":"    fi"},{"line_number":138,"context_line":"}"},{"line_number":139,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"8218d0c1_8ae8e1a4","line":136,"in_reply_to":"5f5dddaa_89b8a89c","updated":"2023-05-24 13:24:09.000000000","message":"So here in the case where we have two nodes, controller for Ceph and subnode for Manila etc. the pre-install and install phases will run using cephadm on the controller. The install ceph config stuff is in the local.conf in that node. In the subnode during devstack install, only the post-config should run and the ceph config options that set up those services are in that local.conf. Post-config will enable services, add pools, create keys and after I scp /etc/ceph to the subnode (I\u0027m not sure where this happens previously as Dan mentioned), configures clients. \n\nI\u0027ve just realized that I\u0027m still running pre-install in the subnode, I\u0027ll have to change that I think. I havent been able to run this whole thing locally because I dont have the correct fsid/cluster plumbed into post-config running on my subnode, so it stops in set_min_client_version for cinder. To fix this, I just need a way to get the cluster id from the controller (The first line from start_ceph will suffice.) I\u0027ll admit I\u0027ve gotten a little confused about client_config, I see theres a ceph.conf being made there, which would override what I got from the scp I did previously. Another option is to have the FSID be part of the local.conf similar to what I tried with SSH_USER (this helps for local deployments mostly). I think much of the confusion here might be because I tried to make this in a way that would work both locally and through CI.\n\nIf we only want to run ceph in one node then all the ceph config options would be in the controller node (including the service configs like manila) BUT I still need to get the client keys and such to the manila node so it can communicate with the cluster. Does the ansible layer copy the /etc/ceph contents or is that somewhere else? Would only copying that file suffice?","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"68306ddff5589a15ab52858425967720a3083d15","unresolved":false,"context_lines":[{"line_number":133,"context_line":"    echo \"[GET CEPHADM] cephadm is now available\""},{"line_number":134,"context_line":""},{"line_number":135,"context_line":"    if [ -z \"$CEPHADM\" ]; then"},{"line_number":136,"context_line":"        CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""},{"line_number":137,"context_line":"    fi"},{"line_number":138,"context_line":"}"},{"line_number":139,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"9bbaac91_bb574302","line":136,"in_reply_to":"679a11a7_6c73b143","updated":"2023-07-24 19:07:16.000000000","message":"Done","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"86652c9608f66ed6258ea92d9de8e93460c38d34","unresolved":true,"context_lines":[{"line_number":133,"context_line":"    echo \"[GET CEPHADM] cephadm is now available\""},{"line_number":134,"context_line":""},{"line_number":135,"context_line":"    if [ -z \"$CEPHADM\" ]; then"},{"line_number":136,"context_line":"        CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""},{"line_number":137,"context_line":"    fi"},{"line_number":138,"context_line":"}"},{"line_number":139,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"bd2d5950_f42148b1","line":136,"in_reply_to":"6bc4cf2a_93dbc1b2","updated":"2023-05-23 20:45:05.000000000","message":"To be fair, I think this is *currently* ansible-ized (at least in upstream CI) so hacking in the ssh stuff is a bit of a regression.\n\nI think that in general just the ceph config and keys are required to be on the subnode, so ssh\u0027ing between them to sudo-run things is a somewhat large jump. Could we not pre-generate that for both nodes in a way that makes it fully parallelizable or reproducible in a smaller environment for people?\n\nEither way, I\u0027m very much in favor of moving to more ansible for this stuff, not less.","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"7cbe0dbc5648869b24f3116237c7d0d11124ec38","unresolved":true,"context_lines":[{"line_number":133,"context_line":"    echo \"[GET CEPHADM] cephadm is now available\""},{"line_number":134,"context_line":""},{"line_number":135,"context_line":"    if [ -z \"$CEPHADM\" ]; then"},{"line_number":136,"context_line":"        CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""},{"line_number":137,"context_line":"    fi"},{"line_number":138,"context_line":"}"},{"line_number":139,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"679a11a7_6c73b143","line":136,"in_reply_to":"8218d0c1_8ae8e1a4","updated":"2023-05-24 13:29:58.000000000","message":"\u003e Does the ansible layer copy the /etc/ceph contents or is that somewhere else? Would only copying that file suffice?\n\nYeah, as I linked below, it copies `/etc/ceph` to the subnode:\n\nhttps://opendev.org/openstack/devstack/src/branch/master/roles/sync-controller-ceph-conf-and-keys/tasks/main.yaml\n\nThat should have the ceph config and keys. Here\u0027s an example from an existing multinode ceph run of what gets copied to the subnode:\n\nhttps://6a4a4acdbde8f30a98f4-d9273944e714c02206b4a053d7e2acce.ssl.cf5.rackcdn.com/879682/6/check/nova-live-migration-ceph/73342c7/compute1/logs/ceph/\n\nSince that will expire, let me confirm for posterity that the directory listing includes `ceph.conf` and `ceph.client.*.keyring` files.","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"14ee66e568fafe295cc79aa136829e31069b878b","unresolved":true,"context_lines":[{"line_number":133,"context_line":"    echo \"[GET CEPHADM] cephadm is now available\""},{"line_number":134,"context_line":""},{"line_number":135,"context_line":"    if [ -z \"$CEPHADM\" ]; then"},{"line_number":136,"context_line":"        CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""},{"line_number":137,"context_line":"    fi"},{"line_number":138,"context_line":"}"},{"line_number":139,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"e7bdde55_9d73f8a1","line":136,"in_reply_to":"b976f754_1c80ca65","updated":"2023-05-23 17:02:22.000000000","message":"In the case that remote_ceph is true, the command to get to cephadm shell is \"ssh $SSH_USER@$CEPH_IP sudo ${TARGET_BIN}/cephadm\" and that is what im using as $CEPHADM. I need sudo to be part of the command othewise I cant reuse CEPHADM as the situation calls. good thing is that cephadm is /always/ called with sudo so its an easy thing to do.","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"27d12ae7fd276b214e401119e340968b570f1908","unresolved":true,"context_lines":[{"line_number":133,"context_line":"    echo \"[GET CEPHADM] cephadm is now available\""},{"line_number":134,"context_line":""},{"line_number":135,"context_line":"    if [ -z \"$CEPHADM\" ]; then"},{"line_number":136,"context_line":"        CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""},{"line_number":137,"context_line":"    fi"},{"line_number":138,"context_line":"}"},{"line_number":139,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"1145e893_12be3511","line":136,"in_reply_to":"bd2d5950_f42148b1","updated":"2023-05-23 21:36:07.000000000","message":"hmmm, i see. As an alternative (because i\u0027m still indexing ansiblizing this by breaking up and running the stack.sh script in stages), would it be okay to just run cephadm from the host locally? we were avoiding installing the binary on all the nodes (ashley, keep me honest - this should work)","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"50cd047f5387348148e7781b6627d8d89b19de82","unresolved":true,"context_lines":[{"line_number":133,"context_line":"    echo \"[GET CEPHADM] cephadm is now available\""},{"line_number":134,"context_line":""},{"line_number":135,"context_line":"    if [ -z \"$CEPHADM\" ]; then"},{"line_number":136,"context_line":"        CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""},{"line_number":137,"context_line":"    fi"},{"line_number":138,"context_line":"}"},{"line_number":139,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"24cf2225_bc832786","line":136,"in_reply_to":"e7bdde55_9d73f8a1","updated":"2023-05-23 17:32:30.000000000","message":"I think you should explicitly put the sudo in the ssh command. However, I still think sshing between the nodes like this is less ideal than letting ansible copy the things that we need (like has been the case for this job for a while).\n\nWhy do you need to ssh to the main node to run `cephadm` commands? Before we just needed ceph config and keys (right?)... what else needs to get run?\n\nPersonally I think that wrapping cephadm in an ssh-to-main-node is likely to be confusing.","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"a427d4c612c2aaf1b2fbd178749db59c8e34d68e","unresolved":true,"context_lines":[{"line_number":507,"context_line":"    bootstrap_config"},{"line_number":508,"context_line":"    get_cephadm"},{"line_number":509,"context_line":"    start_ceph"},{"line_number":510,"context_line":"    sudo cat \u003c\u003cEOF \u003e $CEPH_CONF_DIR/CEPH_CLUSTER_IS_READY.txt"},{"line_number":511,"context_line":"Cluster is provisioned."},{"line_number":512,"context_line":"EOF"},{"line_number":513,"context_line":"}"}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"ab39a40c_0d25f0bf","line":510,"updated":"2023-05-23 14:19:03.000000000","message":"Running cat as sudo will not allow it to write to `$CEPH_CONF_DIR` because the file is actually opened by the shell. You want something like:\n```\ncat \u003c\u003cEOF | sudo tee $CEPH_CONF_DIR/$CEPH_CLUSTER_IS_READY.txt\n...\n```\nHowever since you only seem to care that it\u0027s created, you might as well just:\n```\nsudo touch $CEPH_CONF_DIR/$CEPH_CLUSTER_IS_READY.txt\n```","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"14ee66e568fafe295cc79aa136829e31069b878b","unresolved":true,"context_lines":[{"line_number":507,"context_line":"    bootstrap_config"},{"line_number":508,"context_line":"    get_cephadm"},{"line_number":509,"context_line":"    start_ceph"},{"line_number":510,"context_line":"    sudo cat \u003c\u003cEOF \u003e $CEPH_CONF_DIR/CEPH_CLUSTER_IS_READY.txt"},{"line_number":511,"context_line":"Cluster is provisioned."},{"line_number":512,"context_line":"EOF"},{"line_number":513,"context_line":"}"}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"dc2c6e72_ce133342","line":510,"in_reply_to":"ab39a40c_0d25f0bf","updated":"2023-05-23 17:02:22.000000000","message":"ah yes this makes sense, thanks. Havent worked with bash script before this, so I\u0027m learning as  go","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"f9d6ae915b480f7615313fd75846cce16d273312","unresolved":false,"context_lines":[{"line_number":507,"context_line":"    bootstrap_config"},{"line_number":508,"context_line":"    get_cephadm"},{"line_number":509,"context_line":"    start_ceph"},{"line_number":510,"context_line":"    sudo cat \u003c\u003cEOF \u003e $CEPH_CONF_DIR/CEPH_CLUSTER_IS_READY.txt"},{"line_number":511,"context_line":"Cluster is provisioned."},{"line_number":512,"context_line":"EOF"},{"line_number":513,"context_line":"}"}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"1ae96d86_68983334","line":510,"in_reply_to":"dc2c6e72_ce133342","updated":"2023-05-24 13:24:09.000000000","message":"Done","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"a427d4c612c2aaf1b2fbd178749db59c8e34d68e","unresolved":true,"context_lines":[{"line_number":512,"context_line":"EOF"},{"line_number":513,"context_line":"}"},{"line_number":514,"context_line":""},{"line_number":515,"context_line":"function ceph_is_ready {"},{"line_number":516,"context_line":"    echo \"Waiting the cluster to be up\""},{"line_number":517,"context_line":"    until ssh $SSH_USER@$CEPH_IP ls /etc/ceph/CEPH_CLUSTER_IS_READY.txt \u0026\u003e /dev/null; do"},{"line_number":518,"context_line":"        sleep 1"}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"72f57bfb_0c379175","line":515,"updated":"2023-05-23 14:19:03.000000000","message":"Where is this called from?\n\nI assume it\u0027s going to be from the subnode to the master one? I\u0027m not sure this polling over ssh is really the best way to do it, but I\u0027d have to ask.","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"68306ddff5589a15ab52858425967720a3083d15","unresolved":false,"context_lines":[{"line_number":512,"context_line":"EOF"},{"line_number":513,"context_line":"}"},{"line_number":514,"context_line":""},{"line_number":515,"context_line":"function ceph_is_ready {"},{"line_number":516,"context_line":"    echo \"Waiting the cluster to be up\""},{"line_number":517,"context_line":"    until ssh $SSH_USER@$CEPH_IP ls /etc/ceph/CEPH_CLUSTER_IS_READY.txt \u0026\u003e /dev/null; do"},{"line_number":518,"context_line":"        sleep 1"}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"0066e3c6_253b3f2d","line":515,"in_reply_to":"2e17fb88_513b31db","updated":"2023-07-24 19:07:16.000000000","message":"Done","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"50cd047f5387348148e7781b6627d8d89b19de82","unresolved":true,"context_lines":[{"line_number":512,"context_line":"EOF"},{"line_number":513,"context_line":"}"},{"line_number":514,"context_line":""},{"line_number":515,"context_line":"function ceph_is_ready {"},{"line_number":516,"context_line":"    echo \"Waiting the cluster to be up\""},{"line_number":517,"context_line":"    until ssh $SSH_USER@$CEPH_IP ls /etc/ceph/CEPH_CLUSTER_IS_READY.txt \u0026\u003e /dev/null; do"},{"line_number":518,"context_line":"        sleep 1"}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"2e17fb88_513b31db","line":515,"in_reply_to":"4d83021d_800b9904","updated":"2023-05-23 17:32:30.000000000","message":"Yeah I understand that\u0027s what this code inside the function is doing. What I was saying is I don\u0027t see that it actually gets called anywhere. However, I see the call is in `plugin.sh`.","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"c7f0da6407c575ef34e06a80ea9aab052027514a","unresolved":true,"context_lines":[{"line_number":512,"context_line":"EOF"},{"line_number":513,"context_line":"}"},{"line_number":514,"context_line":""},{"line_number":515,"context_line":"function ceph_is_ready {"},{"line_number":516,"context_line":"    echo \"Waiting the cluster to be up\""},{"line_number":517,"context_line":"    until ssh $SSH_USER@$CEPH_IP ls /etc/ceph/CEPH_CLUSTER_IS_READY.txt \u0026\u003e /dev/null; do"},{"line_number":518,"context_line":"        sleep 1"}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"897ee047_856587a5","line":515,"in_reply_to":"72f57bfb_0c379175","updated":"2023-05-23 14:46:02.000000000","message":"Yeah, so I think we already have a way to make sure the ceph.conf gets copied from the master node to the subnode:\n\nhttps://opendev.org/openstack/devstack/src/branch/master/roles/sync-controller-ceph-conf-and-keys/tasks/main.yaml\n\nAFAIK, that\u0027s how it happens for the current package-based case. Doing this manually in this code is likely to be less reliable I think.","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"14ee66e568fafe295cc79aa136829e31069b878b","unresolved":true,"context_lines":[{"line_number":512,"context_line":"EOF"},{"line_number":513,"context_line":"}"},{"line_number":514,"context_line":""},{"line_number":515,"context_line":"function ceph_is_ready {"},{"line_number":516,"context_line":"    echo \"Waiting the cluster to be up\""},{"line_number":517,"context_line":"    until ssh $SSH_USER@$CEPH_IP ls /etc/ceph/CEPH_CLUSTER_IS_READY.txt \u0026\u003e /dev/null; do"},{"line_number":518,"context_line":"        sleep 1"}],"source_content_type":"application/x-shellscript","patch_set":25,"id":"4d83021d_800b9904","line":515,"in_reply_to":"897ee047_856587a5","updated":"2023-05-23 17:02:22.000000000","message":"So the thought process here is that when the master and subnode are created in parallel like in the case of zuul ci, I need to make sure the node with manila (subnode) waits until ceph is ready on master. this is more to have a waiter than to copy ceph.conf. though that does happen later in post-config","commit_id":"94b582261cc795407188a0479439f36715210691"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"d3d4316049786147a989ab8429d153c119460f6b","unresolved":true,"context_lines":[{"line_number":670,"context_line":"function cleanup_ceph {"},{"line_number":671,"context_line":"    # Cleanup the service."},{"line_number":672,"context_line":"    if [[ \"$REMOTE_CEPH\" \u003d\u003d \"True\" ]]; then"},{"line_number":673,"context_line":"        echo \"Remote Ceph cluster, skip stop_ceph and delete_osd_dev\""},{"line_number":674,"context_line":"    else"},{"line_number":675,"context_line":"        stop_ceph"},{"line_number":676,"context_line":"        delete_osd_dev"}],"source_content_type":"application/x-shellscript","patch_set":43,"id":"cedee1a2_3147386e","line":673,"range":{"start_line":673,"start_character":35,"end_line":673,"end_character":39},"updated":"2023-07-20 23:54:29.000000000","message":"nit: skipping","commit_id":"c4486eb7b79a9a15c5b80d301ebd23681d9d4030"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"68306ddff5589a15ab52858425967720a3083d15","unresolved":false,"context_lines":[{"line_number":670,"context_line":"function cleanup_ceph {"},{"line_number":671,"context_line":"    # Cleanup the service."},{"line_number":672,"context_line":"    if [[ \"$REMOTE_CEPH\" \u003d\u003d \"True\" ]]; then"},{"line_number":673,"context_line":"        echo \"Remote Ceph cluster, skip stop_ceph and delete_osd_dev\""},{"line_number":674,"context_line":"    else"},{"line_number":675,"context_line":"        stop_ceph"},{"line_number":676,"context_line":"        delete_osd_dev"}],"source_content_type":"application/x-shellscript","patch_set":43,"id":"e6459c10_ace4fed6","line":673,"range":{"start_line":673,"start_character":35,"end_line":673,"end_character":39},"in_reply_to":"cedee1a2_3147386e","updated":"2023-07-24 19:07:16.000000000","message":"Done","commit_id":"c4486eb7b79a9a15c5b80d301ebd23681d9d4030"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"9d9509dcabaa7851e3b5c2b6511f47c4709e8a7e","unresolved":true,"context_lines":[{"line_number":98,"context_line":"function set_debug {"},{"line_number":99,"context_line":"    if [ \"$DEBUG\" -eq 1 ]; then"},{"line_number":100,"context_line":"        echo \"[CEPHADM] Enabling Debug mode\""},{"line_number":101,"context_line":"        $SUDO \"$CEPHADM\" shell --fsid $FSID --config $CEPH_CONFIG \\"},{"line_number":102,"context_line":"            --keyring $CEPH_KEYRING -- ceph config set mgr mgr/cephadm/log_to_cluster_level debug"},{"line_number":103,"context_line":"        echo \"[CEPHADM] See debug logs running: ceph -W cephadm --watch-debug\""},{"line_number":104,"context_line":"    fi"}],"source_content_type":"application/x-shellscript","patch_set":46,"id":"cfadf365_497f5753","side":"PARENT","line":101,"updated":"2023-07-26 19:45:33.000000000","message":"As I mentioned before, I\u0027d prefer not to build the sudo into the command variable itself so it\u0027s clear when and where we\u0027re running this with sudo. Since this is already the case, making this change just clouds the rest of the change here. Even if we\u0027re going to do this, I\u0027d rather it be part of a cleanup patch than one that changes the way this all works.","commit_id":"bdce6ce508671b9fb6be7825c8c244e742895362"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b681f164ec6123dc86767b7ca20734823d01dc22","unresolved":false,"context_lines":[{"line_number":98,"context_line":"function set_debug {"},{"line_number":99,"context_line":"    if [ \"$DEBUG\" -eq 1 ]; then"},{"line_number":100,"context_line":"        echo \"[CEPHADM] Enabling Debug mode\""},{"line_number":101,"context_line":"        $SUDO \"$CEPHADM\" shell --fsid $FSID --config $CEPH_CONFIG \\"},{"line_number":102,"context_line":"            --keyring $CEPH_KEYRING -- ceph config set mgr mgr/cephadm/log_to_cluster_level debug"},{"line_number":103,"context_line":"        echo \"[CEPHADM] See debug logs running: ceph -W cephadm --watch-debug\""},{"line_number":104,"context_line":"    fi"}],"source_content_type":"application/x-shellscript","patch_set":46,"id":"c2e3611b_46d6ff9e","side":"PARENT","line":101,"in_reply_to":"cfadf365_497f5753","updated":"2023-08-07 20:11:14.000000000","message":"Done","commit_id":"bdce6ce508671b9fb6be7825c8c244e742895362"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"9d9509dcabaa7851e3b5c2b6511f47c4709e8a7e","unresolved":true,"context_lines":[{"line_number":124,"context_line":"# Pre-install ceph: install required dependencies"},{"line_number":125,"context_line":"function install_deps {"},{"line_number":126,"context_line":"    if [[ \"$REMOTE_CEPH\" \u003d\u003d \"True\" ]]; then"},{"line_number":127,"context_line":"        install_package jq ceph-common"},{"line_number":128,"context_line":"    else"},{"line_number":129,"context_line":"        install_package jq ceph-common"},{"line_number":130,"context_line":"        install_package python3-cephfs python3-prettytable python3-rados python3-rbd python3-requests"}],"source_content_type":"application/x-shellscript","patch_set":46,"id":"0b1ef466_5e81dc48","line":127,"updated":"2023-07-26 19:45:33.000000000","message":"This is minor, but can\u0027t we do this above the if for both cases and only do the python deps if REMOTE_CEPH\u003d\u003dFalse? It\u0027ll just eliminate the duplicate package list for future maintenance.","commit_id":"442ff1d7b0ce6347a45cbac247b34213557b710d"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"d347f282af0929e0e55462278a71c6392eac2fd3","unresolved":true,"context_lines":[{"line_number":124,"context_line":"# Pre-install ceph: install required dependencies"},{"line_number":125,"context_line":"function install_deps {"},{"line_number":126,"context_line":"    if [[ \"$REMOTE_CEPH\" \u003d\u003d \"True\" ]]; then"},{"line_number":127,"context_line":"        install_package jq ceph-common"},{"line_number":128,"context_line":"    else"},{"line_number":129,"context_line":"        install_package jq ceph-common"},{"line_number":130,"context_line":"        install_package python3-cephfs python3-prettytable python3-rados python3-rbd python3-requests"}],"source_content_type":"application/x-shellscript","patch_set":46,"id":"8de316dd_3d27db43","line":127,"in_reply_to":"0b1ef466_5e81dc48","updated":"2023-07-26 19:47:26.000000000","message":"And also, why aren\u0027t these just in the package files like podman et al?","commit_id":"442ff1d7b0ce6347a45cbac247b34213557b710d"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b681f164ec6123dc86767b7ca20734823d01dc22","unresolved":false,"context_lines":[{"line_number":124,"context_line":"# Pre-install ceph: install required dependencies"},{"line_number":125,"context_line":"function install_deps {"},{"line_number":126,"context_line":"    if [[ \"$REMOTE_CEPH\" \u003d\u003d \"True\" ]]; then"},{"line_number":127,"context_line":"        install_package jq ceph-common"},{"line_number":128,"context_line":"    else"},{"line_number":129,"context_line":"        install_package jq ceph-common"},{"line_number":130,"context_line":"        install_package python3-cephfs python3-prettytable python3-rados python3-rbd python3-requests"}],"source_content_type":"application/x-shellscript","patch_set":46,"id":"583efe9f_d1d2cdf8","line":127,"in_reply_to":"868555bb_a16c2573","updated":"2023-08-07 20:11:14.000000000","message":"Done","commit_id":"442ff1d7b0ce6347a45cbac247b34213557b710d"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"3189562ee15077c5936b1b759b3b87670a2da874","unresolved":true,"context_lines":[{"line_number":124,"context_line":"# Pre-install ceph: install required dependencies"},{"line_number":125,"context_line":"function install_deps {"},{"line_number":126,"context_line":"    if [[ \"$REMOTE_CEPH\" \u003d\u003d \"True\" ]]; then"},{"line_number":127,"context_line":"        install_package jq ceph-common"},{"line_number":128,"context_line":"    else"},{"line_number":129,"context_line":"        install_package jq ceph-common"},{"line_number":130,"context_line":"        install_package python3-cephfs python3-prettytable python3-rados python3-rbd python3-requests"}],"source_content_type":"application/x-shellscript","patch_set":46,"id":"868555bb_a16c2573","line":127,"in_reply_to":"8de316dd_3d27db43","updated":"2023-07-26 20:13:28.000000000","message":"+1 ack; i\u0027d prefer that too\n\nWill let Ashley pursue these changes ..","commit_id":"442ff1d7b0ce6347a45cbac247b34213557b710d"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"9d9509dcabaa7851e3b5c2b6511f47c4709e8a7e","unresolved":true,"context_lines":[{"line_number":649,"context_line":""},{"line_number":650,"context_line":"    [ \"$ENABLE_CEPH_RGW\" \u003d\u003d \"True\" ] \u0026\u0026 SERVICES+\u003d(\u0027rgw\u0027)"},{"line_number":651,"context_line":""},{"line_number":652,"context_line":""},{"line_number":653,"context_line":"    enable_services"},{"line_number":654,"context_line":"    add_pools"},{"line_number":655,"context_line":"    create_keys"}],"source_content_type":"application/x-shellscript","patch_set":46,"id":"b4a49e81_11dcd02e","line":652,"updated":"2023-07-26 19:45:33.000000000","message":"Unrelated whitespace damage","commit_id":"442ff1d7b0ce6347a45cbac247b34213557b710d"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b681f164ec6123dc86767b7ca20734823d01dc22","unresolved":false,"context_lines":[{"line_number":649,"context_line":""},{"line_number":650,"context_line":"    [ \"$ENABLE_CEPH_RGW\" \u003d\u003d \"True\" ] \u0026\u0026 SERVICES+\u003d(\u0027rgw\u0027)"},{"line_number":651,"context_line":""},{"line_number":652,"context_line":""},{"line_number":653,"context_line":"    enable_services"},{"line_number":654,"context_line":"    add_pools"},{"line_number":655,"context_line":"    create_keys"}],"source_content_type":"application/x-shellscript","patch_set":46,"id":"447cde06_8e87c500","line":652,"in_reply_to":"b4a49e81_11dcd02e","updated":"2023-08-07 20:11:14.000000000","message":"Done","commit_id":"442ff1d7b0ce6347a45cbac247b34213557b710d"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"05f5e39baa8cf342ab3ca59bd95b57770896eeab","unresolved":true,"context_lines":[{"line_number":289,"context_line":"    $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID --config $CEPH_CONFIG \\"},{"line_number":290,"context_line":"        --keyring $CEPH_KEYRING -- ceph auth get-or-create \"$name\" mgr \"allow rw\" mon \"allow r\" osd \"$osd_caps\" \\"},{"line_number":291,"context_line":"        -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\" || $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID \\"},{"line_number":292,"context_line":"        --config $CEPH_CONFIG --keyring $CEPH_KEYRING -- ceph auth get \"$name\" -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\""},{"line_number":293,"context_line":""},{"line_number":294,"context_line":""},{"line_number":295,"context_line":"    $SUDO chown ${STACK_USER}:$(id -g -n $whoami) \\"}],"source_content_type":"application/x-shellscript","patch_set":50,"id":"8c984f41_5f4d645d","line":292,"updated":"2023-08-22 21:44:00.000000000","message":"Hmm, the only difference here is \"get_or_create\" vs \"get\" (other than the default perms). Is there some reason why we need to do this? If so, can you add a comment about why? This is a pretty (darn) thick compound command here so it\u0027d be good to know why we need to do this.\n\nPerhaps it\u0027s because the remote machine may reach this point first and it would create the key before the main machine would? Even still, I\u0027d expect `get_or_create` to work as it describes.","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"126ec5e95f384c380493d5edb54b1c95134280a5","unresolved":true,"context_lines":[{"line_number":289,"context_line":"    $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID --config $CEPH_CONFIG \\"},{"line_number":290,"context_line":"        --keyring $CEPH_KEYRING -- ceph auth get-or-create \"$name\" mgr \"allow rw\" mon \"allow r\" osd \"$osd_caps\" \\"},{"line_number":291,"context_line":"        -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\" || $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID \\"},{"line_number":292,"context_line":"        --config $CEPH_CONFIG --keyring $CEPH_KEYRING -- ceph auth get \"$name\" -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\""},{"line_number":293,"context_line":""},{"line_number":294,"context_line":""},{"line_number":295,"context_line":"    $SUDO chown ${STACK_USER}:$(id -g -n $whoami) \\"}],"source_content_type":"application/x-shellscript","patch_set":50,"id":"a4486ecc_1f0b6794","line":292,"in_reply_to":"4246a6ff_deab3389","updated":"2023-08-24 13:54:37.000000000","message":"Yep, agree, it would be better to make it explicit which node does that work. I originally had written something similar but couldn\u0027t resolve why `get_or_create` wouldn\u0027t *technically* work. However, better to be explicit and deterministic, IMHO, unless there\u0027s some reason we can\u0027t.","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":25402,"name":"Francesco Pantano","email":"fpantano@redhat.com","username":"fmount"},"change_message_id":"53554ac23c3a1f26f1c454fd37c6b82e7667f43b","unresolved":true,"context_lines":[{"line_number":289,"context_line":"    $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID --config $CEPH_CONFIG \\"},{"line_number":290,"context_line":"        --keyring $CEPH_KEYRING -- ceph auth get-or-create \"$name\" mgr \"allow rw\" mon \"allow r\" osd \"$osd_caps\" \\"},{"line_number":291,"context_line":"        -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\" || $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID \\"},{"line_number":292,"context_line":"        --config $CEPH_CONFIG --keyring $CEPH_KEYRING -- ceph auth get \"$name\" -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\""},{"line_number":293,"context_line":""},{"line_number":294,"context_line":""},{"line_number":295,"context_line":"    $SUDO chown ${STACK_USER}:$(id -g -n $whoami) \\"}],"source_content_type":"application/x-shellscript","patch_set":50,"id":"4246a6ff_deab3389","line":292,"in_reply_to":"8c984f41_5f4d645d","updated":"2023-08-24 07:57:33.000000000","message":"Hi Dan/Asheley,\nI think in general, in a multinode context (but even when the Ceph cluster\nis external), the `create_keys` function should be run under a `REMOTE_CEPH` `if`\ncondition to reflect the following:\n\n1. `REMOTE_CEPH \u003d false`: I\u0027m deploying the Ceph cluster (and I actually call the\n`bootstrap` function), hence I should create the keys according to the enabled_services\n\n2. `REMOTE_CEPH \u003d True`: my cluster is already deployed elsewhere, ceph.conf and keyrings\nare supposed to be copied from the \"ansible layer\", I also get the FSID by running\n`cat $CEPH_CONFIG ...` on L37; I have all the information in place, and I think\nwe have no reason to regenerate keys with wrong info.\nFor similar reasons, by simply calling \"get\" we\u0027re trying to \"export\" in the\ntarget location keyring information that are already present in the node.\n\nUnless we only copy ceph.conf omitting keyrings (which I don\u0027t think so given we\u0027re\npassing the admin keyring to this function), L300 should be in a form like:\n\n```\nif [[ \"$REMOTE_CEPH\" \u003d \"False\" ]]; then\n    add_pools\n    create_keys\n```\nand this function restored to the original content.","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"cd096f596b6254a61ea03c90ee3b48161e341b27","unresolved":true,"context_lines":[{"line_number":289,"context_line":"    $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID --config $CEPH_CONFIG \\"},{"line_number":290,"context_line":"        --keyring $CEPH_KEYRING -- ceph auth get-or-create \"$name\" mgr \"allow rw\" mon \"allow r\" osd \"$osd_caps\" \\"},{"line_number":291,"context_line":"        -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\" || $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID \\"},{"line_number":292,"context_line":"        --config $CEPH_CONFIG --keyring $CEPH_KEYRING -- ceph auth get \"$name\" -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\""},{"line_number":293,"context_line":""},{"line_number":294,"context_line":""},{"line_number":295,"context_line":"    $SUDO chown ${STACK_USER}:$(id -g -n $whoami) \\"}],"source_content_type":"application/x-shellscript","patch_set":50,"id":"78e74e8f_763dba63","line":292,"in_reply_to":"a4486ecc_1f0b6794","updated":"2023-09-07 18:03:04.000000000","message":"ack;","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"cd096f596b6254a61ea03c90ee3b48161e341b27","unresolved":true,"context_lines":[{"line_number":289,"context_line":"    $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID --config $CEPH_CONFIG \\"},{"line_number":290,"context_line":"        --keyring $CEPH_KEYRING -- ceph auth get-or-create \"$name\" mgr \"allow rw\" mon \"allow r\" osd \"$osd_caps\" \\"},{"line_number":291,"context_line":"        -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\" || $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID \\"},{"line_number":292,"context_line":"        --config $CEPH_CONFIG --keyring $CEPH_KEYRING -- ceph auth get \"$name\" -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\""},{"line_number":293,"context_line":""},{"line_number":294,"context_line":""},{"line_number":295,"context_line":"    $SUDO chown ${STACK_USER}:$(id -g -n $whoami) \\"}],"source_content_type":"application/x-shellscript","patch_set":50,"id":"ef707e46_898bcbad","line":292,"in_reply_to":"a4486ecc_1f0b6794","updated":"2023-09-07 18:03:04.000000000","message":"ack; the get_or_create was failing because of \"missing\" OSD caps.. so the logic here was incorrect in a sense.. so i\u0027m okay with us dropping this and relying on the ansible layer to handle the conf and keyring copy in the CI jobs... on local devstacks, we should probably add a note that the /etc/ceph directory must contain the necessary config and keys..","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"fee6eca750c0820f60e70bfc62cbd5e1a236170b","unresolved":false,"context_lines":[{"line_number":289,"context_line":"    $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID --config $CEPH_CONFIG \\"},{"line_number":290,"context_line":"        --keyring $CEPH_KEYRING -- ceph auth get-or-create \"$name\" mgr \"allow rw\" mon \"allow r\" osd \"$osd_caps\" \\"},{"line_number":291,"context_line":"        -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\" || $SUDO \"$CEPHADM\" shell -v \"$KEY_EXPORT_DIR:$KEY_EXPORT_DIR\" --fsid $FSID \\"},{"line_number":292,"context_line":"        --config $CEPH_CONFIG --keyring $CEPH_KEYRING -- ceph auth get \"$name\" -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\""},{"line_number":293,"context_line":""},{"line_number":294,"context_line":""},{"line_number":295,"context_line":"    $SUDO chown ${STACK_USER}:$(id -g -n $whoami) \\"}],"source_content_type":"application/x-shellscript","patch_set":50,"id":"f0c12f8c_fce5755c","line":292,"in_reply_to":"ef707e46_898bcbad","updated":"2023-09-12 20:52:48.000000000","message":"Ack","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":25402,"name":"Francesco Pantano","email":"fpantano@redhat.com","username":"fmount"},"change_message_id":"53554ac23c3a1f26f1c454fd37c6b82e7667f43b","unresolved":true,"context_lines":[{"line_number":651,"context_line":""},{"line_number":652,"context_line":"    enable_services"},{"line_number":653,"context_line":"    add_pools"},{"line_number":654,"context_line":"    create_keys"},{"line_number":655,"context_line":"    client_config"},{"line_number":656,"context_line":"    import_libvirt_secret_ceph"},{"line_number":657,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":50,"id":"b5e5446c_cfcaec00","line":654,"updated":"2023-08-24 07:57:33.000000000","message":"```\nif [[ \"$REMOTE_CEPH\" \u003d \"False\" ]]; then\n    add_pools\n    create_keys\nfi\n```","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"fee6eca750c0820f60e70bfc62cbd5e1a236170b","unresolved":false,"context_lines":[{"line_number":651,"context_line":""},{"line_number":652,"context_line":"    enable_services"},{"line_number":653,"context_line":"    add_pools"},{"line_number":654,"context_line":"    create_keys"},{"line_number":655,"context_line":"    client_config"},{"line_number":656,"context_line":"    import_libvirt_secret_ceph"},{"line_number":657,"context_line":""}],"source_content_type":"application/x-shellscript","patch_set":50,"id":"fd9e6851_10725810","line":654,"in_reply_to":"b5e5446c_cfcaec00","updated":"2023-09-12 20:52:48.000000000","message":"Done","commit_id":"8df7e358f8b76b905ca3dbca43e9dba454d8f1ff"},{"author":{"_account_id":29632,"name":"Carlos Eduardo","email":"ces.eduardo98@gmail.com","username":"silvacarlos"},"change_message_id":"132d667b4ea6f69d1a0667a4a46d113f0bcddb2c","unresolved":true,"context_lines":[{"line_number":287,"context_line":"        --keyring $CEPH_KEYRING -- ceph auth get-or-create \"$name\" mgr \"allow rw\" mon \"allow r\" osd \"$osd_caps\" \\"},{"line_number":288,"context_line":"        -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\""},{"line_number":289,"context_line":""},{"line_number":290,"context_line":""},{"line_number":291,"context_line":"    $SUDO chown ${STACK_USER}:$(id -g -n $whoami) \\"},{"line_number":292,"context_line":"        ${CEPH_CONF_DIR}/ceph.$name.keyring"},{"line_number":293,"context_line":"}"}],"source_content_type":"application/x-shellscript","patch_set":54,"id":"6f395abc_74ce4413","line":290,"updated":"2023-09-15 17:07:43.000000000","message":"nit: any reason for the extra blank line?","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b9d5bdd7cbbd7f32997b2a98a12d77ff188d1a46","unresolved":true,"context_lines":[{"line_number":287,"context_line":"        --keyring $CEPH_KEYRING -- ceph auth get-or-create \"$name\" mgr \"allow rw\" mon \"allow r\" osd \"$osd_caps\" \\"},{"line_number":288,"context_line":"        -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\""},{"line_number":289,"context_line":""},{"line_number":290,"context_line":""},{"line_number":291,"context_line":"    $SUDO chown ${STACK_USER}:$(id -g -n $whoami) \\"},{"line_number":292,"context_line":"        ${CEPH_CONF_DIR}/ceph.$name.keyring"},{"line_number":293,"context_line":"}"}],"source_content_type":"application/x-shellscript","patch_set":54,"id":"3a9a41f4_4ad3eda3","line":290,"in_reply_to":"6a4af345_cb762fc6","updated":"2023-09-18 16:50:49.000000000","message":"Agreed, must be a leftover from a previous edit. Priority right now is to get this patch merged as there is a manila-tempest-plugin job that depends on it. I can submit a follow up to clean this up.","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"def1b5c98a27957aba8dd2c9cee0e72350821c25","unresolved":true,"context_lines":[{"line_number":287,"context_line":"        --keyring $CEPH_KEYRING -- ceph auth get-or-create \"$name\" mgr \"allow rw\" mon \"allow r\" osd \"$osd_caps\" \\"},{"line_number":288,"context_line":"        -o \"$KEY_EXPORT_DIR/ceph.$name.keyring\""},{"line_number":289,"context_line":""},{"line_number":290,"context_line":""},{"line_number":291,"context_line":"    $SUDO chown ${STACK_USER}:$(id -g -n $whoami) \\"},{"line_number":292,"context_line":"        ${CEPH_CONF_DIR}/ceph.$name.keyring"},{"line_number":293,"context_line":"}"}],"source_content_type":"application/x-shellscript","patch_set":54,"id":"6a4af345_cb762fc6","line":290,"in_reply_to":"6f395abc_74ce4413","updated":"2023-09-18 16:38:16.000000000","message":"++ This should be avoided as it makes things look like they were changed in the commit history, when they weren\u0027t.","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"}],"devstack/plugin.sh":[{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"39632bb79a1f42a981fc231a75cc8f546fb773eb","unresolved":true,"context_lines":[{"line_number":40,"context_line":"        fi"},{"line_number":41,"context_line":"    fi"},{"line_number":42,"context_line":"elif [[ \"$1\" \u003d\u003d \"stack\" \u0026\u0026 \"$2\" \u003d\u003d \"install\" ]]; then"},{"line_number":43,"context_line":"    if [[ \"$CEPHADM_DEPLOY\" \u003d \"True\" \u0026\u0026 \"$REMOTE_CEPH\" \u003d \"False\" ]]; then"},{"line_number":44,"context_line":"        # Perform installation of service source"},{"line_number":45,"context_line":"        echo_summary \"[cephadm] Installing ceph\""},{"line_number":46,"context_line":"        install_ceph"}],"source_content_type":"text/x-sh","patch_set":19,"id":"6899cbb2_a50d22d9","line":43,"range":{"start_line":43,"start_character":40,"end_line":43,"end_character":64},"updated":"2023-05-19 00:52:55.000000000","message":"This variable, \"REMOTE_CEPH\" doesn\u0027t have a default; so the script is just skipping onto the \"else\" part here (lines 59-60): https://zuul.opendev.org/t/openstack/build/ddfeeaa802cb4be0bef46aab36ce114e/log/job-output.txt#11800\n\nYou need to set the default for this in the settings file or the override_defaults file: https://github.com/openstack/devstack-plugin-ceph/blob/master/devstack/override-defaults","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b85a4b484c1faab9cb54db6445a734e8935848a5","unresolved":false,"context_lines":[{"line_number":40,"context_line":"        fi"},{"line_number":41,"context_line":"    fi"},{"line_number":42,"context_line":"elif [[ \"$1\" \u003d\u003d \"stack\" \u0026\u0026 \"$2\" \u003d\u003d \"install\" ]]; then"},{"line_number":43,"context_line":"    if [[ \"$CEPHADM_DEPLOY\" \u003d \"True\" \u0026\u0026 \"$REMOTE_CEPH\" \u003d \"False\" ]]; then"},{"line_number":44,"context_line":"        # Perform installation of service source"},{"line_number":45,"context_line":"        echo_summary \"[cephadm] Installing ceph\""},{"line_number":46,"context_line":"        install_ceph"}],"source_content_type":"text/x-sh","patch_set":19,"id":"b189feec_0f8cf30e","line":43,"range":{"start_line":43,"start_character":40,"end_line":43,"end_character":64},"in_reply_to":"6899cbb2_a50d22d9","updated":"2023-05-19 19:58:59.000000000","message":"Done","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b85a4b484c1faab9cb54db6445a734e8935848a5","unresolved":false,"context_lines":[{"line_number":40,"context_line":"        fi"},{"line_number":41,"context_line":"    fi"},{"line_number":42,"context_line":"elif [[ \"$1\" \u003d\u003d \"stack\" \u0026\u0026 \"$2\" \u003d\u003d \"install\" ]]; then"},{"line_number":43,"context_line":"    if [[ \"$CEPHADM_DEPLOY\" \u003d \"True\" \u0026\u0026 \"$REMOTE_CEPH\" \u003d \"False\" ]]; then"},{"line_number":44,"context_line":"        # Perform installation of service source"},{"line_number":45,"context_line":"        echo_summary \"[cephadm] Installing ceph\""},{"line_number":46,"context_line":"        install_ceph"}],"source_content_type":"text/x-sh","patch_set":19,"id":"b276c96a_961f2bfe","line":43,"range":{"start_line":43,"start_character":40,"end_line":43,"end_character":64},"in_reply_to":"6899cbb2_a50d22d9","updated":"2023-05-19 19:58:59.000000000","message":"Done","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":16643,"name":"Goutham Pacha Ravi","email":"gouthampravi@gmail.com","username":"gouthamr"},"change_message_id":"39632bb79a1f42a981fc231a75cc8f546fb773eb","unresolved":true,"context_lines":[{"line_number":62,"context_line":"elif [[ \"$1\" \u003d\u003d \"stack\" \u0026\u0026 \"$2\" \u003d\u003d \"post-config\" ]]; then"},{"line_number":63,"context_line":"    if [[ \"$CEPHADM_DEPLOY\" \u003d \"True\" ]]; then"},{"line_number":64,"context_line":"        if [[ \"$REMOTE_CEPH\" \u003d \"True\" ]]; then"},{"line_number":65,"context_line":"            export CEPHADM\u003d\"ssh stack@$CEPH_IP sudo ${TARGET_BIN}/cephadm\""},{"line_number":66,"context_line":"            ceph_is_ready"},{"line_number":67,"context_line":"        else"},{"line_number":68,"context_line":"            export CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""}],"source_content_type":"text/x-sh","patch_set":19,"id":"9d0d1c32_099e3df7","line":65,"range":{"start_line":65,"start_character":32,"end_line":65,"end_character":37},"updated":"2023-05-19 00:52:55.000000000","message":"same comment here as [1], the SSH user must be configurable.. \n\n[1] https://review.opendev.org/c/openstack/devstack-plugin-ceph/+/876747/19/devstack/lib/cephadm#661","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b85a4b484c1faab9cb54db6445a734e8935848a5","unresolved":false,"context_lines":[{"line_number":62,"context_line":"elif [[ \"$1\" \u003d\u003d \"stack\" \u0026\u0026 \"$2\" \u003d\u003d \"post-config\" ]]; then"},{"line_number":63,"context_line":"    if [[ \"$CEPHADM_DEPLOY\" \u003d \"True\" ]]; then"},{"line_number":64,"context_line":"        if [[ \"$REMOTE_CEPH\" \u003d \"True\" ]]; then"},{"line_number":65,"context_line":"            export CEPHADM\u003d\"ssh stack@$CEPH_IP sudo ${TARGET_BIN}/cephadm\""},{"line_number":66,"context_line":"            ceph_is_ready"},{"line_number":67,"context_line":"        else"},{"line_number":68,"context_line":"            export CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""}],"source_content_type":"text/x-sh","patch_set":19,"id":"2b29b263_43bc1725","line":65,"range":{"start_line":65,"start_character":32,"end_line":65,"end_character":37},"in_reply_to":"9d0d1c32_099e3df7","updated":"2023-05-19 19:58:59.000000000","message":"Done","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b85a4b484c1faab9cb54db6445a734e8935848a5","unresolved":false,"context_lines":[{"line_number":62,"context_line":"elif [[ \"$1\" \u003d\u003d \"stack\" \u0026\u0026 \"$2\" \u003d\u003d \"post-config\" ]]; then"},{"line_number":63,"context_line":"    if [[ \"$CEPHADM_DEPLOY\" \u003d \"True\" ]]; then"},{"line_number":64,"context_line":"        if [[ \"$REMOTE_CEPH\" \u003d \"True\" ]]; then"},{"line_number":65,"context_line":"            export CEPHADM\u003d\"ssh stack@$CEPH_IP sudo ${TARGET_BIN}/cephadm\""},{"line_number":66,"context_line":"            ceph_is_ready"},{"line_number":67,"context_line":"        else"},{"line_number":68,"context_line":"            export CEPHADM\u003d\"${SUDO} ${TARGET_BIN}/cephadm\""}],"source_content_type":"text/x-sh","patch_set":19,"id":"47693a89_e4c3c2a9","line":65,"range":{"start_line":65,"start_character":32,"end_line":65,"end_character":37},"in_reply_to":"9d0d1c32_099e3df7","updated":"2023-05-19 19:58:59.000000000","message":"Done","commit_id":"bb523234a4fcbd4ee02a8427b28bc176d32a3efc"},{"author":{"_account_id":25402,"name":"Francesco Pantano","email":"fpantano@redhat.com","username":"fmount"},"change_message_id":"102db741b4a600a57936a104a830ad2035f4cd74","unresolved":true,"context_lines":[{"line_number":44,"context_line":"        # Perform installation of service source"},{"line_number":45,"context_line":"        echo_summary \"[cephadm] Installing ceph\""},{"line_number":46,"context_line":"        install_ceph"},{"line_number":47,"context_line":"        set_min_client_version"},{"line_number":48,"context_line":"    elif [[ \"$CEPHADM_DEPLOY\" \u003d \"True\" \u0026\u0026 \"$REMOTE_CEPH\" \u003d \"True\" ]]; then"},{"line_number":49,"context_line":"        echo \"[CEPHADM] Remote Ceph: Skipping install\""},{"line_number":50,"context_line":"        get_cephadm"}],"source_content_type":"text/x-sh","patch_set":54,"id":"45ec0206_e94c79f9","line":47,"updated":"2023-09-15 18:37:24.000000000","message":"any reason to call this function here instead of keeping it within `install_ceph`?\nPreviously we used to set `min_client_version` for cinder, but I understand it\u0027s something generic enough and valid for every rbd client.\nBeing that said, considering this is an option set globally in Ceph but for OSDs, I\u0027d rather move this definition here [1], so that we can set it after the Ceph bootstrap.\nCalling it here reaches the same behavior though, so it\u0027s really a minor thing and I was mostly curious about this choice.\n\n[1] https://github.com/openstack/devstack-plugin-ceph/blob/master/devstack/lib/cephadm#L528","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"def1b5c98a27957aba8dd2c9cee0e72350821c25","unresolved":true,"context_lines":[{"line_number":44,"context_line":"        # Perform installation of service source"},{"line_number":45,"context_line":"        echo_summary \"[cephadm] Installing ceph\""},{"line_number":46,"context_line":"        install_ceph"},{"line_number":47,"context_line":"        set_min_client_version"},{"line_number":48,"context_line":"    elif [[ \"$CEPHADM_DEPLOY\" \u003d \"True\" \u0026\u0026 \"$REMOTE_CEPH\" \u003d \"True\" ]]; then"},{"line_number":49,"context_line":"        echo \"[CEPHADM] Remote Ceph: Skipping install\""},{"line_number":50,"context_line":"        get_cephadm"}],"source_content_type":"text/x-sh","patch_set":54,"id":"cca05fce_eb4bea95","line":47,"in_reply_to":"45ec0206_e94c79f9","updated":"2023-09-18 16:38:16.000000000","message":"Was there an answer to this?","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"},{"author":{"_account_id":32594,"name":"Ashley Rodriguez","email":"ashrod98@redhat.com","username":"ashrod98"},"change_message_id":"b9d5bdd7cbbd7f32997b2a98a12d77ff188d1a46","unresolved":true,"context_lines":[{"line_number":44,"context_line":"        # Perform installation of service source"},{"line_number":45,"context_line":"        echo_summary \"[cephadm] Installing ceph\""},{"line_number":46,"context_line":"        install_ceph"},{"line_number":47,"context_line":"        set_min_client_version"},{"line_number":48,"context_line":"    elif [[ \"$CEPHADM_DEPLOY\" \u003d \"True\" \u0026\u0026 \"$REMOTE_CEPH\" \u003d \"True\" ]]; then"},{"line_number":49,"context_line":"        echo \"[CEPHADM] Remote Ceph: Skipping install\""},{"line_number":50,"context_line":"        get_cephadm"}],"source_content_type":"text/x-sh","patch_set":54,"id":"9465f481_a10281f4","line":47,"in_reply_to":"cca05fce_eb4bea95","updated":"2023-09-18 16:50:49.000000000","message":"there was no particular reason to call it here instead of install_ceph, like francesco said it acheives the same behavior. I can address this in a follow up.","commit_id":"b663a9fb838ae6632a3b7e30ee3b4a36a88454f5"}]}
