Commit a2d1cd02 by Czémán Arnold

fix CephDisk.download(), remove CephDisk.merge_disk_with_base(), fix exists() and make_free_space()

parent 3746c537
...@@ -553,7 +553,9 @@ class CephDisk(Disk): ...@@ -553,7 +553,9 @@ class CephDisk(Disk):
with CephConnection(self.dir) as conn: with CephConnection(self.dir) as conn:
rbd_inst = rbd.RBD() rbd_inst = rbd.RBD()
# keep calm, Ceph Block Device uses thin-provisioning # keep calm, Ceph Block Device uses thin-provisioning
rbd_inst.create(conn.ioctx, self.name, int(MAXIMUM_SIZE)) rbd_inst.create(conn.ioctx, self.name, int(MAXIMUM_SIZE),
old_format=False,
features=rbd.RBD_FEATURE_LAYERING)
with rbd.Image(conn.ioctx, self.name) as image: with rbd.Image(conn.ioctx, self.name) as image:
offset = 0 offset = 0
actsize = 0 actsize = 0
...@@ -633,28 +635,6 @@ class CephDisk(Disk): ...@@ -633,28 +635,6 @@ class CephDisk(Disk):
self.__with_ceph_connection(self.__snapshot) self.__with_ceph_connection(self.__snapshot)
def merge_disk_with_base(self, ioctx, task, new_disk, parent_id=None):
with rbd.Image(ioctx, self.name) as image:
logger.debug("Merging %s into %s.",
self.get_path(),
new_disk.get_path())
image.create_snap(new_disk.name)
image.protect_snap(new_disk.name)
if not task.is_aborted():
task.update_state(task_id=parent_id,
state=task.AsyncResult(parent_id).state,
meta={'size': new_disk.size, 'percent': 100})
else:
logger.warning("Merging new disk %s is aborted by user.",
new_disk.get_path())
logger.warning("Aborted merge job, removing %s",
new_disk.get_path())
with rbd.Image(ioctx, self.name) as image:
image.remove_snap(new_disk.name)
def merge_disk_without_base(self, ioctx, task, new_disk, parent_id=None, def merge_disk_without_base(self, ioctx, task, new_disk, parent_id=None,
length=1024 * 1024): length=1024 * 1024):
...@@ -688,9 +668,6 @@ class CephDisk(Disk): ...@@ -688,9 +668,6 @@ class CephDisk(Disk):
if task.is_aborted(): if task.is_aborted():
raise AbortException() raise AbortException()
# if self.base_name:
# self.merge_disk_with_base(ioctx, task, new_disk, parent_id)
# else:
self.merge_disk_without_base(ioctx, task, new_disk, parent_id) self.merge_disk_without_base(ioctx, task, new_disk, parent_id)
def merge(self, task, new_disk, parent_id=None): def merge(self, task, new_disk, parent_id=None):
......
...@@ -62,7 +62,6 @@ def delete(disk_desc): ...@@ -62,7 +62,6 @@ def delete(disk_desc):
@celery.task() @celery.task()
def delete_dump(data_store_type, dir, filename): def delete_dump(data_store_type, dir, filename):
if data_store_type == "ceph_block": if data_store_type == "ceph_block":
with CephConnection(str(dir)) as conn: with CephConnection(str(dir)) as conn:
rbd_inst = rbd.RBD() rbd_inst = rbd.RBD()
...@@ -135,7 +134,7 @@ def exists(data_store_type, path, disk_name): ...@@ -135,7 +134,7 @@ def exists(data_store_type, path, disk_name):
if data_store_type == "ceph_block": if data_store_type == "ceph_block":
try: try:
with CephConnection(str(path)) as conn: with CephConnection(str(path)) as conn:
with rbd.Image(conn.ioctx, disk_name): with rbd.Image(conn.ioctx, str(disk_name)):
pass pass
except rbd.ImageNotFound: except rbd.ImageNotFound:
return False return False
...@@ -163,10 +162,16 @@ def make_free_space(data_store_type, path, deletable_disks, percent=10): ...@@ -163,10 +162,16 @@ def make_free_space(data_store_type, path, deletable_disks, percent=10):
if ds_type == "ceph_block": if ds_type == "ceph_block":
with CephConnection(str(path)) as conn: with CephConnection(str(path)) as conn:
rbd_inst = rbd.RBD() rbd_inst = rbd.RBD()
rbd_inst.remove(conn.ioctx, f) with rbd.Image(conn.ioctx, str(f)) as image:
for snapshot in image.list_snaps():
name = snapshot["name"]
image.unprotect_snap(name)
image.remove_snap(name)
rbd_inst.remove(conn.ioctx, str(f))
else: else:
unlink(os.path.join(path, f)) unlink(os.path.join(path, f))
logger.info('Image: %s removed.' % f) logger.info('Image: %s removed.' % f)
except IndexError: except IndexError:
raise Exception("Has no deletable disk.") logger.warning("Has no deletable disk.")
return False
return True return True
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or sign in to comment