)]}'
{"nova/compute/resource_tracker.py":[{"author":{"_account_id":6873,"name":"Matt Riedemann","email":"mriedem.os@gmail.com","username":"mriedem"},"change_message_id":"47704177ae25c2c5969420ab582e6d822abf1a43","unresolved":false,"context_lines":[{"line_number":785,"context_line":"                        # being deleted or shelved offloaded."},{"line_number":786,"context_line":"                        instance.vm_state not in"},{"line_number":787,"context_line":"                        vm_states.ALLOW_RESOURCE_REMOVAL):"},{"line_number":788,"context_line":"                    LOG.info(\u0027Re-fetching instances since at least one (%s) \u0027"},{"line_number":789,"context_line":"                             \u0027with vm_state (%s) and task_state (%s) is\u0027"},{"line_number":790,"context_line":"                             \u0027tracked against this host (%s) in the \u0027"},{"line_number":791,"context_line":"                             \u0027database but not in the ResourceTracker.\u0027,"}],"source_content_type":"text/x-python","patch_set":1,"id":"7faddb67_4dcd0ab6","line":788,"updated":"2019-08-22 15:22:41.000000000","message":"This is showing up in grenade jobs which makes sense when we\u0027re starting up on the new side, tracked_instances will be empty but there would be instances on the host:\n\nhttp://logstash.openstack.org/#dashboard/file/logstash.json?query\u003dmessage%3A%5C%22Re-fetching%20instances%5C%22%20AND%20tags%3A%5C%22screen-n-cpu.txt%5C%22\u0026from\u003d7d\n\nSo we should avoid this if startup is True.","commit_id":"9b549a434919791cc42f4f92862be964ab5ee164"},{"author":{"_account_id":11564,"name":"Chris Dent","email":"cdent@anticdent.org","username":"chdent"},"change_message_id":"d30d9673a61f8ea708f8a634b358abe899fd989a","unresolved":false,"context_lines":[{"line_number":776,"context_line":"            # not add the instance to tracked_instances again, and"},{"line_number":777,"context_line":"            # _update_usage_from_instance shouldn\u0027t for the deleted case but"},{"line_number":778,"context_line":"            # I\u0027m not sure about the migration case, so that could be a bug."},{"line_number":779,"context_line":"            for instance in instances:"},{"line_number":780,"context_line":"                # TODO(mriedem): Should we check if the instance uuid is in"},{"line_number":781,"context_line":"                # tracked_migrations meaning the instance is moving off or"},{"line_number":782,"context_line":"                # onto this host?"}],"source_content_type":"text/x-python","patch_set":2,"id":"7faddb67_dbfcfb40","line":779,"updated":"2019-08-28 13:35:09.000000000","message":"What you\u0027ve done here is probably more explicit, and since it does a break (below) we don\u0027t have to traverse the whole list if there is a mismatch, however, if we believe that the common case is that we _won\u0027t_ break, then we are doing a repeated \u0027instance.uuid in self.tracked_instances\u0027.\n\nWe could instead\n\n    instance_uuids \u003d set([instance.uuid for instance in instances])\n    # if the sets aren\u0027t concurrent load again\n    set_diff \u003d instance_uuids - self.tracked_instances\n    if set_diff:\n        if check vm_state of the set_diff:\n            do the reload\n\nor some such. May not be worth it, but I\u0027ve become very conscious of the cost of big loops in nova-compute lately...","commit_id":"f1c14bb89fc158069fcaa0adde7db497cac67cfb"},{"author":{"_account_id":6873,"name":"Matt Riedemann","email":"mriedem.os@gmail.com","username":"mriedem"},"change_message_id":"b6fd38e21eef698af086fd5169d3a0b944640708","unresolved":false,"context_lines":[{"line_number":785,"context_line":"                        # being deleted or shelved offloaded."},{"line_number":786,"context_line":"                        instance.vm_state not in"},{"line_number":787,"context_line":"                        vm_states.ALLOW_RESOURCE_REMOVAL):"},{"line_number":788,"context_line":"                    LOG.info(\u0027Re-fetching instances since at least one (%s) \u0027"},{"line_number":789,"context_line":"                             \u0027with vm_state (%s) and task_state (%s) is\u0027"},{"line_number":790,"context_line":"                             \u0027tracked against this host (%s) in the \u0027"},{"line_number":791,"context_line":"                             \u0027database but not in the ResourceTracker.\u0027,"}],"source_content_type":"text/x-python","patch_set":2,"id":"7faddb67_a0cfcd89","line":788,"updated":"2019-08-23 15:46:55.000000000","message":"Still hitting these log statements in PS2, though it seems to have cutout more of the grenade jobs (the startup condition that is), but hitting it in live migration runs:\n\nhttp://logstash.openstack.org/#dashboard/file/logstash.json?query\u003dmessage%3A%5C%22Re-fetching%20instances%5C%22%20AND%20tags%3A%5C%22screen-n-cpu.txt%5C%22%20AND%20build_patchset%3A%5C%222%5C%22\u0026from\u003d7d\n\nBut it\u0027s weird because the task_state is None:\n\n[None req-e60a18fe-c9b2-4e91-a1fb-8d66cae9eaf5 None None] Re-fetching instances since at least one (1837710a-3e11-4f44-93fb-a65653feaaa3) with vm_state (active) and task_state (None) istracked against this host (ubuntu-bionic-rax-ord-0010373580) in the database but not in the ResourceTracker.\n\nI haven\u0027t dug into what\u0027s going on with that instance when this is logged though.","commit_id":"f1c14bb89fc158069fcaa0adde7db497cac67cfb"},{"author":{"_account_id":6873,"name":"Matt Riedemann","email":"mriedem.os@gmail.com","username":"mriedem"},"change_message_id":"b6fd38e21eef698af086fd5169d3a0b944640708","unresolved":false,"context_lines":[{"line_number":786,"context_line":"                        instance.vm_state not in"},{"line_number":787,"context_line":"                        vm_states.ALLOW_RESOURCE_REMOVAL):"},{"line_number":788,"context_line":"                    LOG.info(\u0027Re-fetching instances since at least one (%s) \u0027"},{"line_number":789,"context_line":"                             \u0027with vm_state (%s) and task_state (%s) is\u0027"},{"line_number":790,"context_line":"                             \u0027tracked against this host (%s) in the \u0027"},{"line_number":791,"context_line":"                             \u0027database but not in the ResourceTracker.\u0027,"},{"line_number":792,"context_line":"                             instance.uuid, instance.vm_state,"}],"source_content_type":"text/x-python","patch_set":2,"id":"7faddb67_c0d44922","line":789,"range":{"start_line":789,"start_character":71,"end_line":789,"end_character":72},"updated":"2019-08-23 15:46:55.000000000","message":"space","commit_id":"f1c14bb89fc158069fcaa0adde7db497cac67cfb"},{"author":{"_account_id":11564,"name":"Chris Dent","email":"cdent@anticdent.org","username":"chdent"},"change_message_id":"d30d9673a61f8ea708f8a634b358abe899fd989a","unresolved":false,"context_lines":[{"line_number":795,"context_line":"                        context, nodename)"},{"line_number":796,"context_line":"                    # Sanity check to see if this even helped."},{"line_number":797,"context_line":"                    # TODO(mriedem): Should likely compare UUIDs vs length."},{"line_number":798,"context_line":"                    if len(instances) \u003d\u003d len(new_instances):"},{"line_number":799,"context_line":"                        LOG.warning(\u0027Re-fetching instances retrieved the same \u0027"},{"line_number":800,"context_line":"                                    \u0027number of instances. Was it worth it?\u0027)"},{"line_number":801,"context_line":"                    instances \u003d new_instances"}],"source_content_type":"text/x-python","patch_set":2,"id":"7faddb67_1b3cd3df","line":798,"updated":"2019-08-28 13:35:09.000000000","message":"same set stuff might make sense here, would also give you something interesting to put in the log message","commit_id":"f1c14bb89fc158069fcaa0adde7db497cac67cfb"}]}
