)]}'
{"/PATCHSET_LEVEL":[{"author":{"_account_id":5202,"name":"Erno Kuvaja","email":"jokke@usr.fi","username":"jokke"},"change_message_id":"83a5eea2feeaef773bfbec50e98eca96177fd772","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"629aa6b9_a372230b","updated":"2024-06-07 08:25:42.000000000","message":"This is highlighting the issue I have been flagging around our gabbi tests. Tempest is meant to be able to run against real operational clouds. Someone without great understanding of the telemetry-tempest-plugin internals could run this suite with real credentials. The gabbi tests would fail on the very initial test because it doesn\u0027t get empty stack list, never create any resources, but this cleanup would kill all the stacks that the user has running potentially leading to data loss.\n\nThe only way to do this safely is if we can capture the resources gabbi run is creating and only clean up those, not everything we see under the credentials.","commit_id":"2496d69a073a9417dff76b1b03ffdfe9e8d9d580"},{"author":{"_account_id":5202,"name":"Erno Kuvaja","email":"jokke@usr.fi","username":"jokke"},"change_message_id":"3004d1925e1c2f6ea8f0387b7038bc663cbdc317","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"496f5c63_f7bd94ee","in_reply_to":"2f59cb36_2cd60093","updated":"2024-06-07 09:00:57.000000000","message":"https://docs.openstack.org/tempest/latest/configuration.html See pre-provisioned credentials section. Also note, while the section does mention that the project should be empty or the tempest run might fail, it does not say that tempest might wipe all your resources.","commit_id":"2496d69a073a9417dff76b1b03ffdfe9e8d9d580"},{"author":{"_account_id":34975,"name":"Jaromír Wysoglad","email":"jwysogla@redhat.com","username":"jwysogla"},"change_message_id":"e238bd56dbf029eafe66f0fb9497fc38dedc33a3","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"29ab8c2e_7b191aa5","in_reply_to":"496f5c63_f7bd94ee","updated":"2024-06-07 12:26:45.000000000","message":"I moved the stack name generation out of _prep_test() to resource_setup(). The generated name is saved into a variable and the proposed cleanup code now looks for a stack with exactly that name. Example stack name generated during my testing: \"tempest-telemetry-1807709925\"","commit_id":"2496d69a073a9417dff76b1b03ffdfe9e8d9d580"},{"author":{"_account_id":34975,"name":"Jaromír Wysoglad","email":"jwysogla@redhat.com","username":"jwysogla"},"change_message_id":"67f051d3dd4b0ae8b612de6b107d4c66af9ea73b","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"2f59cb36_2cd60093","in_reply_to":"629aa6b9_a372230b","updated":"2024-06-07 08:41:21.000000000","message":"Can the user run the tests with real credentials? In my understanding the ScenarioTest will automatically create (and at the end of the tests remove) the \"admin\" and \"primary\" credentials. Could you point me to how I could run tempest with real credentials, so that I can test this situation?","commit_id":"2496d69a073a9417dff76b1b03ffdfe9e8d9d580"},{"author":{"_account_id":5202,"name":"Erno Kuvaja","email":"jokke@usr.fi","username":"jokke"},"change_message_id":"a72175288e9e94148cc29973f297845bde4ebe2e","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"1d29dd6c_d960e53b","updated":"2024-06-07 14:15:21.000000000","message":"Addresses my concern, now just to figure out why Zuul doesn\u0027t like it","commit_id":"b33ad02061238f1026a605bea8593eca2c1dc923"},{"author":{"_account_id":34975,"name":"Jaromír Wysoglad","email":"jwysogla@redhat.com","username":"jwysogla"},"change_message_id":"610ce138707aa9f6a39c12288451df404d7009e3","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"59bb307c_704f2d2c","updated":"2024-06-10 08:55:18.000000000","message":"I created a new patch, which adds polling for the deleted servers, which should hopefully get rid of the issue we see in the CI here: https://review.opendev.org/c/openstack/telemetry-tempest-plugin/+/921646","commit_id":"b33ad02061238f1026a605bea8593eca2c1dc923"},{"author":{"_account_id":34975,"name":"Jaromír Wysoglad","email":"jwysogla@redhat.com","username":"jwysogla"},"change_message_id":"e791ff59d9971f6cfadf1b9f4bc67409df93c1a3","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"6c9b3223_ce5eda93","updated":"2024-06-10 08:31:05.000000000","message":"recheck","commit_id":"b33ad02061238f1026a605bea8593eca2c1dc923"},{"author":{"_account_id":34975,"name":"Jaromír Wysoglad","email":"jwysogla@redhat.com","username":"jwysogla"},"change_message_id":"460a82c0a7703306c4353ef018c4dfe0b3b7d7c6","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"5ef400b4_c16c23fb","updated":"2024-06-10 05:27:15.000000000","message":"recheck - based on the logs it looks like a server was still running during the last step of the autoscaling scenario. We might want to poll multiple times and wait a little for everything to get deleted. This doesn\u0027t seem related to this patch.","commit_id":"b33ad02061238f1026a605bea8593eca2c1dc923"},{"author":{"_account_id":34975,"name":"Jaromír Wysoglad","email":"jwysogla@redhat.com","username":"jwysogla"},"change_message_id":"1a42788105d2cdda2855e1b715017d5fd7e1c5e9","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"b596ac4a_dad729a0","updated":"2024-06-10 08:31:28.000000000","message":"same issue as before","commit_id":"b33ad02061238f1026a605bea8593eca2c1dc923"},{"author":{"_account_id":4264,"name":"Matthias Runge","email":"mrunge@redhat.com","username":"mrunge"},"change_message_id":"dfd5557401ce8a18482ace57fa653c8289215cd8","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":2,"id":"61f32bd3_37055707","updated":"2024-06-20 06:55:10.000000000","message":"thank you!","commit_id":"b33ad02061238f1026a605bea8593eca2c1dc923"}]}
