--- /dev/null Thu Jan 01 00:00:00 1970 +0000
+++ b/components/openstack/nova/patches/09-CVE-2015-3280.patch Tue Oct 20 05:38:12 2015 -0700
@@ -0,0 +1,200 @@
+Upstream patch to address CVE-2015-3280. This fix will be included in
+the future 2014.2.3 (juno) release.
+
+From fa72fb8b51d59e04913c871539cee98a3da79058 Mon Sep 17 00:00:00 2001
+From: Rajesh Tailor <[email protected]>
+Date: Wed, 4 Mar 2015 05:05:19 -0800
+Subject: Delete orphaned instance files from compute nodes
+
+While resizing/revert-resizing instance, if instance gets deleted
+in between, then instance files remains either on the source or
+destination compute node.
+
+To address this issue, added a new periodic task
+'_cleanup_incomplete_migrations' which takes care of deleting
+instance files from source/destination compute nodes and then
+mark migration record as failed so that it doesn't appear again
+in the next periodic task run.
+
+Conflicts:
+ nova/compute/manager.py
+ nova/tests/unit/compute/test_compute_mgr.py
+
+SecurityImpact
+Closes-Bug: 1392527
+Change-Id: I9866d8e32e99b9f907921f4b226edf7b62bd83a7
+(cherry picked from commit 18d6b5cc79973fc553daf7a92f22cce4dc0ca013)
+
+--- nova-2014.2.2/nova/compute/manager.py.~1~ 2015-09-02 14:46:43.532548379 -0700
++++ nova-2014.2.2/nova/compute/manager.py 2015-09-02 14:47:57.813280934 -0700
+@@ -257,12 +257,18 @@ def errors_out_migration(function):
+ def decorated_function(self, context, *args, **kwargs):
+ try:
+ return function(self, context, *args, **kwargs)
+- except Exception:
++ except Exception as ex:
+ with excutils.save_and_reraise_exception():
+ migration = kwargs['migration']
+- status = migration.status
+- if status not in ['migrating', 'post-migrating']:
+- return
++
++ # NOTE(rajesht): If InstanceNotFound error is thrown from
++ # decorated function, migration status should be set to
++ # 'error', without checking current migration status.
++ if not isinstance(ex, exception.InstanceNotFound):
++ status = migration.status
++ if status not in ['migrating', 'post-migrating']:
++ return
++
+ migration.status = 'error'
+ try:
+ migration.save(context.elevated())
+@@ -3469,6 +3475,7 @@ class ComputeManager(manager.Manager):
+ @wrap_exception()
+ @reverts_task_state
+ @wrap_instance_event
++ @errors_out_migration
+ @wrap_instance_fault
+ def revert_resize(self, context, instance, migration, reservations):
+ """Destroys the new instance on the destination machine.
+@@ -3523,6 +3530,7 @@ class ComputeManager(manager.Manager):
+ @wrap_exception()
+ @reverts_task_state
+ @wrap_instance_event
++ @errors_out_migration
+ @wrap_instance_fault
+ def finish_revert_resize(self, context, instance, reservations, migration):
+ """Finishes the second half of reverting a resize.
+@@ -6246,3 +6254,48 @@ class ComputeManager(manager.Manager):
+ instance.cleaned = True
+ with utils.temporary_mutation(context, read_deleted='yes'):
+ instance.save(context)
++
++ @periodic_task.periodic_task(spacing=CONF.instance_delete_interval)
++ def _cleanup_incomplete_migrations(self, context):
++ """Delete instance files on failed resize/revert-resize operation
++
++ During resize/revert-resize operation, if that instance gets deleted
++ in-between then instance files might remain either on source or
++ destination compute node because of race condition.
++ """
++ LOG.debug('Cleaning up deleted instances with incomplete migration ')
++ migration_filters = {'host':CONF.host,
++ 'status': 'error'}
++ migrations = objects.MigrationList.get_by_filters(context,
++ migration_filters)
++
++ if not migrations:
++ return
++
++ inst_uuid_from_migrations = set([migration.instance_uuid for migration
++ in migrations])
++
++ inst_filters = {'deleted': True, 'soft_deleted': False,
++ 'uuid': inst_uuid_from_migrations}
++ attrs = ['info_cache', 'security_groups', 'system_metadata']
++ with utils.temporary_mutation(context, read_deleted='yes'):
++ instances = objects.InstanceList.get_by_filters(
++ context, inst_filters, expected_attrs=attrs, use_slave=True)
++
++ for instance in instances:
++ if instance.host != CONF.host:
++ for migration in migrations:
++ if instance.uuid == migration.instance_uuid:
++ # Delete instance files if not cleanup properly either
++ # from the source or destination compute nodes when
++ # the instance is deleted during resizing.
++ self.driver.delete_instance_files(instance)
++ try:
++ migration.status = 'failed'
++ migration.save(context.elevated())
++ except exception.MigrationNotFound:
++ LOG.warning(_LW("Migration %s is not found."),
++ migration.id, context=context,
++ instance=instance)
++ break
++
+
+--- ./nova/tests/compute/test_compute_mgr.py.~1~ 2015-09-29 09:45:07.760433246 -0700
++++ ./nova/tests/compute/test_compute_mgr.py 2015-09-29 09:48:00.008811912 -0700
+@@ -1047,6 +1047,79 @@ class ComputeManagerUnitTestCase(test.No
+ self.assertFalse(c.cleaned)
+ self.assertEqual('1', c.system_metadata['clean_attempts'])
+
++ @mock.patch.object(objects.Migration, 'save')
++ @mock.patch.object(objects.MigrationList, 'get_by_filters')
++ @mock.patch.object(objects.InstanceList, 'get_by_filters')
++ def _test_cleanup_incomplete_migrations(self, inst_host,
++ mock_inst_get_by_filters,
++ mock_migration_get_by_filters,
++ mock_save):
++ def fake_inst(context, uuid, host):
++ inst = objects.Instance(context)
++ inst.uuid = uuid
++ inst.host = host
++ return inst
++
++ def fake_migration(uuid, status, inst_uuid, src_host, dest_host):
++ migration = objects.Migration()
++ migration.uuid = uuid
++ migration.status = status
++ migration.instance_uuid = inst_uuid
++ migration.source_compute = src_host
++ migration.dest_compute = dest_host
++ return migration
++
++ fake_instances = [fake_inst(self.context, '111', inst_host),
++ fake_inst(self.context, '222', inst_host)]
++
++ fake_migrations = [fake_migration('123', 'error', '111',
++ 'fake-host', 'fake-mini'),
++ fake_migration('456', 'error', '222',
++ 'fake-host', 'fake-mini')]
++
++ mock_migration_get_by_filters.return_value = fake_migrations
++ mock_inst_get_by_filters.return_value = fake_instances
++
++ with mock.patch.object(self.compute.driver, 'delete_instance_files'):
++ self.compute._cleanup_incomplete_migrations(self.context)
++
++ # Ensure that migration status is set to 'failed' after instance
++ # files deletion for those instances whose instance.host is not
++ # same as compute host where periodic task is running.
++ for inst in fake_instances:
++ if inst.host != CONF.host:
++ for mig in fake_migrations:
++ if inst.uuid == mig.instance_uuid:
++ self.assertEqual('failed', mig.status)
++
++ def test_cleanup_incomplete_migrations_dest_node(self):
++ """Test to ensure instance files are deleted from destination node.
++
++ If an instance gets deleted during resizing/revert-resizing
++ operation, in that case instance files gets deleted from
++ instance.host (source host here), but there is possibility that
++ instance files could be present on destination node.
++
++ This test ensures that `_cleanup_incomplete_migration` periodic
++ task deletes orphaned instance files from destination compute node.
++ """
++ self.flags(host='fake-mini')
++ self._test_cleanup_incomplete_migrations('fake-host')
++
++ def test_cleanup_incomplete_migrations_source_node(self):
++ """Test to ensure instance files are deleted from source node.
++
++ If instance gets deleted during resizing/revert-resizing operation,
++ in that case instance files gets deleted from instance.host (dest
++ host here), but there is possibility that instance files could be
++ present on source node.
++
++ This test ensures that `_cleanup_incomplete_migration` periodic
++ task deletes orphaned instance files from source compute node.
++ """
++ self.flags(host='fake-host')
++ self._test_cleanup_incomplete_migrations('fake-mini')
++
+ def test_attach_interface_failure(self):
+ # Test that the fault methods are invoked when an attach fails
+ db_instance = fake_instance.fake_db_instance()