)]}'
{"/COMMIT_MSG":[{"author":{"_account_id":7166,"name":"Sylvain Bauza","email":"sbauza@redhat.com","username":"sbauza"},"change_message_id":"f3311e96b60352dd6fd725de86a465932b2f8504","unresolved":true,"context_lines":[{"line_number":28,"context_line":"on a test-by-test basis, as Cinder has done [1], but that seems more"},{"line_number":29,"context_line":"complicated for no good reason. Yet another alternative would be to"},{"line_number":30,"context_line":"reintroduce the serialization of these tests first introduced in change"},{"line_number":31,"context_line":"I6ce930fa86c82da1008089791942b1fff7d04c18, but that is left until later"},{"line_number":32,"context_line":"in the hopes that simply increasing the timeout will resolve the issues."},{"line_number":33,"context_line":""},{"line_number":34,"context_line":"[1] https://github.com/openstack/cinder/blob/19.0.0/cinder/tests/unit/db/test_migrations.py"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":1,"id":"8b983595_964a80e0","line":31,"updated":"2021-10-22 09:12:41.000000000","message":"as discussed on IRC, this above change modified the behaviour, so that\u0027s why we need to update the timeout.","commit_id":"fb083138eb03ca6de3e90e07a9c603b717e73fbe"}],"/PATCHSET_LEVEL":[{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"dc66daf4cac91c4980563c019bb4a1f0fdb8cc99","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"9b1f93c5_3cfe7b54","updated":"2021-10-19 11:23:59.000000000","message":"Also note that this is already helping the aarch64 jobs. From [1]\n\n  2021-10-19 10:43:30.761988 | ubuntu-focal-arm64 | [5583] /home/zuul/src/opendev.org/openstack/nova$ /home/zuul/src/opendev.org/openstack/nova/.tox/py38/bin/stestr slowest\n  2021-10-19 10:43:40.601110 | ubuntu-focal-arm64 | Test id                                                                                  Runtime (s)\n  2021-10-19 10:43:40.601216 | ubuntu-focal-arm64 | ---------------------------------------------------------------------------------------  -----------\n  2021-10-19 10:43:40.601255 | ubuntu-focal-arm64 | nova.tests.unit.db.main.test_migrations.TestModelsLegacySyncMySQL.test_models_sync       239.773\n  2021-10-19 10:43:40.601290 | ubuntu-focal-arm64 | nova.tests.unit.db.main.test_migrations.TestModelsSyncMySQL.test_models_sync             233.313\n  2021-10-19 10:43:40.601324 | ubuntu-focal-arm64 | nova.tests.unit.db.main.test_migrations.TestModelsSyncMySQL.test_innodb_tables           228.683\n  2021-10-19 10:43:40.601392 | ubuntu-focal-arm64 | nova.tests.unit.db.main.test_migrations.TestMigrationsWalkMySQL.test_walk_versions       225.721\n  2021-10-19 10:43:40.601428 | ubuntu-focal-arm64 | nova.tests.unit.db.api.test_migrations.TestModelsSyncMySQL.test_models_sync               78.743\n  2021-10-19 10:43:40.601456 | ubuntu-focal-arm64 | nova.tests.unit.db.api.test_migrations.TestModelsLegacySyncMySQL.test_models_sync         75.591\n  2021-10-19 10:43:40.601487 | ubuntu-focal-arm64 | nova.tests.unit.db.api.test_migrations.TestMigrationsWalkMySQL.test_walk_versions         69.210\n  2021-10-19 10:43:40.601519 | ubuntu-focal-arm64 | nova.tests.unit.db.main.test_migrations.TestModelsLegacySyncPostgreSQL.test_models_sync   46.141\n  2021-10-19 10:43:40.601548 | ubuntu-focal-arm64 | nova.tests.unit.db.main.test_migrations.TestModelsSyncPostgreSQL.test_models_sync         24.501\n  2021-10-19 10:43:40.601578 | ubuntu-focal-arm64 | nova.tests.unit.db.main.test_migrations.TestMigrationsWalkPostgreSQL.test_walk_versions   21.946\n\nWithout this change, the top 4 of those would have failed. Now we have a good deal of headroom (160 seconds again being the default timeout and 640 seconds being the new one, which is substantially higher than the current highest runtime)\n\n[1] https://zuul.opendev.org/t/openstack/build/a272a21e21ff483c8c584f5948ff5849/log/job-output.txt#19059","commit_id":"fb083138eb03ca6de3e90e07a9c603b717e73fbe"},{"author":{"_account_id":9708,"name":"Balazs Gibizer","display_name":"gibi","email":"gibizer@gmail.com","username":"gibi"},"change_message_id":"d05f743e3170abd7fc7f2a30fe8c1d9b49b49558","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"abf8153b_6c7d198a","updated":"2021-10-19 10:45:09.000000000","message":"Fine by me. ","commit_id":"fb083138eb03ca6de3e90e07a9c603b717e73fbe"},{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"b8f1094e9252754a68349984ab623dc47e4cd71a","unresolved":false,"context_lines":[],"source_content_type":"","patch_set":1,"id":"fd202507_eca723fa","updated":"2021-10-19 11:20:22.000000000","message":"Some comparisons of runtimes across passing and failing CI jobs at [1]. I don\u0027t have a clear grasp yet as to why this is happening but this patch should at least stop the bleeding.\n\n[1] https://paste.opendev.org/show/810073/","commit_id":"fb083138eb03ca6de3e90e07a9c603b717e73fbe"}]}
