Skip to content
Toggle navigation
P
Projects
G
Groups
S
Snippets
Help
CIRCLE
/
vmdriver
This project
Loading...
Sign in
Toggle navigation
Go to a project
Project
Repository
Issues
2
Merge Requests
0
Wiki
Members
Activity
Graph
Charts
Create a new issue
Commits
Issue Boards
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Commit
2d1af23c
authored
Jun 14, 2016
by
Czémán Arnold
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Rework ceph operations with new configuration method
parent
f759eb24
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
90 additions
and
29 deletions
+90
-29
ceph.py
+82
-18
vm.py
+6
-9
vmdriver.py
+2
-2
No files found.
ceph.py
View file @
2d1af23c
...
@@ -6,30 +6,62 @@ import libvirt
...
@@ -6,30 +6,62 @@ import libvirt
import
lxml.etree
as
ET
import
lxml.etree
as
ET
from
base64
import
b64decode
from
base64
import
b64decode
import
logging
import
logging
import
re
import
json
from
util
import
req_connection
,
wrap_libvirtError
,
Connection
from
util
import
req_connection
,
wrap_libvirtError
,
Connection
logger
=
logging
.
getLogger
(
__name__
)
logger
=
logging
.
getLogger
(
__name__
)
DUMP_SIZE_LIMIT
=
int
(
os
.
getenv
(
"DUMP_SIZE_LIMIT"
,
20
*
1024
**
3
))
# 20GB
DUMP_SIZE_LIMIT
=
int
(
os
.
getenv
(
"DUMP_SIZE_LIMIT"
,
20
*
1024
**
3
))
# 20GB
mon_regex
=
re
.
compile
(
r"^\[(?P<address>.+)\]\:(?P<port>\d+).*$"
)
class
CephConfig
:
def
__init__
(
self
,
user
=
None
,
config_path
=
None
,
keyring_path
=
None
):
self
.
user
=
user
self
.
config_path
=
config_path
self
.
keyring_path
=
keyring_path
if
user
is
None
:
self
.
user
=
"admin"
if
config_path
is
None
:
self
.
config_path
=
os
.
getenv
(
"CEPH_CONFIG"
,
"/etc/ceph/ceph.conf"
)
if
keyring_path
is
None
:
default_keyring
=
"/etc/ceph/ceph.client.
%
s.keyring"
%
self
.
user
self
.
keyring_path
=
os
.
getenv
(
"CEPH_KEYRING"
,
default_keyring
)
def
cmd_args
(
self
):
return
[
"--keyring"
,
self
.
keyring_path
,
"--id"
,
self
.
user
,
"--conf"
,
self
.
config_path
]
class
CephConnection
:
class
CephConnection
:
def
__init__
(
self
,
pool_name
,
c
eph_config
=
None
):
def
__init__
(
self
,
pool_name
,
c
onf
=
None
):
self
.
pool_name
=
pool_name
self
.
pool_name
=
pool_name
self
.
ceph_config
=
ceph_config
self
.
conf
=
conf
if
conf
is
None
:
self
.
conf
=
CephConfig
()
self
.
cluster
=
None
self
.
cluster
=
None
self
.
ioctx
=
None
self
.
ioctx
=
None
def
__enter__
(
self
):
def
__enter__
(
self
):
try
:
try
:
if
self
.
ceph_config
is
None
:
self
.
cluster
=
rados
.
Rados
(
self
.
ceph_config
=
os
.
getenv
(
"CEPH_CONFIG"
,
conffile
=
self
.
conf
.
config_path
,
"/etc/ceph/ceph.conf"
)
conf
=
dict
(
keyring
=
self
.
conf
.
keyring_path
)
)
self
.
cluster
=
rados
.
Rados
(
conffile
=
self
.
ceph_config
)
timeout
=
os
.
getenv
(
"CEPH_TIMEOUT"
,
2
)
self
.
cluster
.
connect
(
timeout
=
2
)
self
.
cluster
.
connect
(
timeout
=
timeout
)
self
.
ioctx
=
self
.
cluster
.
open_ioctx
(
self
.
pool_name
)
self
.
ioctx
=
self
.
cluster
.
open_ioctx
(
self
.
pool_name
)
except
rados
.
InterruptedOrTimeoutError
as
e
:
except
rados
.
InterruptedOrTimeoutError
as
e
:
raise
Exception
(
e
)
raise
Exception
(
e
)
...
@@ -46,12 +78,41 @@ def sudo(*args):
...
@@ -46,12 +78,41 @@ def sudo(*args):
subprocess
.
check_output
([
"/bin/sudo"
]
+
list
(
args
))
subprocess
.
check_output
([
"/bin/sudo"
]
+
list
(
args
))
def
map_rbd
(
ceph_path
,
local_path
):
def
unmap_rbd
(
conf
,
local_path
):
sudo
(
"/bin/rbd"
,
"unmap"
,
local_path
,
*
conf
.
cmd_args
())
def
map_rbd
(
conf
,
ceph_path
,
local_path
):
try
:
try
:
sudo
(
"/bin/rbd"
,
"map"
,
ceph_path
)
sudo
(
"/bin/rbd"
,
"map"
,
ceph_path
,
*
conf
.
cmd_args
()
)
except
:
except
:
sudo
(
"/bin/rbd"
,
"unmap"
,
local_path
)
unmap_rbd
(
conf
,
local_path
)
sudo
(
"/bin/rbd"
,
"map"
,
ceph_path
)
sudo
(
"/bin/rbd"
,
"map"
,
ceph_path
,
*
conf
.
cmd_args
())
def
get_secret_key
(
conf
,
user
):
return
subprocess
.
check_output
(([
"/bin/ceph"
,
"auth"
,
"print-key"
,
"client.
%
s"
%
user
]
+
conf
.
cmd_args
()))
def
parse_endpoint
(
mon
):
m
=
mon_regex
.
match
(
mon
[
"addr"
])
return
(
m
.
group
(
"address"
),
m
.
group
(
"port"
))
def
_get_endpoints
(
conf
):
output
=
subprocess
.
check_output
(([
"/bin/ceph"
,
"mon"
,
"dump"
,
"--format=json"
]
+
conf
.
cmd_args
()))
mon_data
=
json
.
loads
(
output
)
mons
=
mon_data
[
"mons"
]
return
map
(
parse_endpoint
,
mons
)
def
get_endpoints
(
user
):
conf
=
CephConfig
(
user
=
user
)
return
_get_endpoints
(
conf
)
def
save
(
domain
,
poolname
,
diskname
):
def
save
(
domain
,
poolname
,
diskname
):
...
@@ -69,13 +130,13 @@ def save(domain, poolname, diskname):
...
@@ -69,13 +130,13 @@ def save(domain, poolname, diskname):
rbd_inst
.
remove
(
conn
.
ioctx
,
diskname
)
rbd_inst
.
remove
(
conn
.
ioctx
,
diskname
)
rbd_inst
.
create
(
conn
.
ioctx
,
diskname
,
disk_size
)
rbd_inst
.
create
(
conn
.
ioctx
,
diskname
,
disk_size
)
try
:
try
:
map_rbd
(
ceph_path
,
local_path
)
map_rbd
(
c
onn
.
conf
,
c
eph_path
,
local_path
)
domain
.
save
(
local_path
)
domain
.
save
(
local_path
)
except
:
except
:
rbd_inst
.
remove
(
conn
.
ioctx
,
diskname
)
rbd_inst
.
remove
(
conn
.
ioctx
,
diskname
)
raise
raise
finally
:
finally
:
sudo
(
"/bin/rbd"
,
"unmap"
,
local_path
)
unmap_rbd
(
conn
.
conf
,
local_path
)
def
restore
(
connection
,
poolname
,
diskname
):
def
restore
(
connection
,
poolname
,
diskname
):
...
@@ -84,9 +145,9 @@ def restore(connection, poolname, diskname):
...
@@ -84,9 +145,9 @@ def restore(connection, poolname, diskname):
ceph_path
=
os
.
path
.
join
(
poolname
,
diskname
)
ceph_path
=
os
.
path
.
join
(
poolname
,
diskname
)
local_path
=
os
.
path
.
join
(
"/dev/rbd"
,
ceph_path
)
local_path
=
os
.
path
.
join
(
"/dev/rbd"
,
ceph_path
)
map_rbd
(
ceph_path
,
local_path
)
map_rbd
(
c
onnection
.
conf
,
c
eph_path
,
local_path
)
connection
.
restore
(
local_path
)
connection
.
restore
(
local_path
)
sudo
(
"/bin/rbd"
,
"unmap"
,
local_path
)
unmap_rbd
(
connection
.
conf
,
local_path
)
with
CephConnection
(
poolname
)
as
conn
:
with
CephConnection
(
poolname
)
as
conn
:
rbd_inst
=
rbd
.
RBD
()
rbd_inst
=
rbd
.
RBD
()
rbd_inst
.
remove
(
conn
.
ioctx
,
diskname
)
rbd_inst
.
remove
(
conn
.
ioctx
,
diskname
)
...
@@ -123,7 +184,10 @@ def find_secret(user):
...
@@ -123,7 +184,10 @@ def find_secret(user):
@req_connection
@req_connection
@wrap_libvirtError
@wrap_libvirtError
def
create_secret
(
user
,
secretkey
):
def
create_secret
(
user
):
conf
=
CephConfig
()
secretkey
=
get_secret_key
(
conf
,
user
)
xml
=
generate_secret_xml
(
user
)
xml
=
generate_secret_xml
(
user
)
conn
=
Connection
.
get
()
conn
=
Connection
.
get
()
secret
=
conn
.
secretDefineXML
(
xml
)
secret
=
conn
.
secretDefineXML
(
xml
)
...
@@ -141,9 +205,9 @@ def delete_secret(user):
...
@@ -141,9 +205,9 @@ def delete_secret(user):
logger
.
info
(
"Secret with uuid: '
%
s' deleted"
,
secret
.
UUIDString
())
logger
.
info
(
"Secret with uuid: '
%
s' deleted"
,
secret
.
UUIDString
())
def
check_secret
(
user
,
secretkey
):
def
check_secret
(
user
):
secret
=
find_secret
(
user
)
secret
=
find_secret
(
user
)
if
secret
is
None
:
if
secret
is
None
:
secret
=
create_secret
(
user
,
secretkey
)
secret
=
create_secret
(
user
)
return
secret
.
UUIDString
()
return
secret
.
UUIDString
()
vm.py
View file @
2d1af23c
import
lxml.etree
as
ET
import
lxml.etree
as
ET
from
vmcelery
import
native_ovs
from
vmcelery
import
native_ovs
from
ceph
import
check_secret
from
ceph
import
check_secret
,
get_endpoints
# VM Instance class
# VM Instance class
...
@@ -246,7 +246,6 @@ class CephVMDisk(VMDisk):
...
@@ -246,7 +246,6 @@ class CephVMDisk(VMDisk):
def
__init__
(
self
,
def
__init__
(
self
,
source
,
source
,
endpoints
,
disk_device
=
"disk"
,
disk_device
=
"disk"
,
driver_name
=
"qemu"
,
driver_name
=
"qemu"
,
driver_type
=
"raw"
,
driver_type
=
"raw"
,
...
@@ -254,8 +253,7 @@ class CephVMDisk(VMDisk):
...
@@ -254,8 +253,7 @@ class CephVMDisk(VMDisk):
target_device
=
"vda"
,
target_device
=
"vda"
,
target_bus
=
"virtio"
,
target_bus
=
"virtio"
,
protocol
=
"rbd"
,
protocol
=
"rbd"
,
ceph_user
=
None
,
ceph_user
=
None
):
secret
=
None
):
super
(
CephVMDisk
,
self
)
.
__init__
(
super
(
CephVMDisk
,
self
)
.
__init__
(
source
=
source
,
source
=
source
,
...
@@ -267,12 +265,11 @@ class CephVMDisk(VMDisk):
...
@@ -267,12 +265,11 @@ class CephVMDisk(VMDisk):
target_device
=
target_device
,
target_device
=
target_device
,
target_bus
=
target_bus
)
target_bus
=
target_bus
)
self
.
endpoints
=
endpoints
self
.
protocol
=
protocol
self
.
protocol
=
protocol
self
.
ceph_user
=
ceph_user
self
.
ceph_user
=
ceph_user
self
.
secret
=
secret
if
ceph_user
is
not
None
:
if
ceph_user
is
not
None
and
secret
is
not
None
:
check_secret
(
ceph_user
)
check_secret
(
ceph_user
,
secret
)
self
.
endpoints
=
get_endpoints
(
ceph_user
)
@classmethod
@classmethod
def
deserialize
(
cls
,
desc
):
def
deserialize
(
cls
,
desc
):
...
@@ -291,7 +288,7 @@ class CephVMDisk(VMDisk):
...
@@ -291,7 +288,7 @@ class CephVMDisk(VMDisk):
ET
.
SubElement
(
source
,
"host"
,
ET
.
SubElement
(
source
,
"host"
,
attrib
=
{
"name"
:
name
,
"port"
:
unicode
(
port
)})
attrib
=
{
"name"
:
name
,
"port"
:
unicode
(
port
)})
if
self
.
ceph_user
is
not
None
and
self
.
secret
is
not
None
:
if
self
.
ceph_user
is
not
None
:
auth
=
ET
.
SubElement
(
auth
=
ET
.
SubElement
(
xml_top
,
xml_top
,
"auth"
,
"auth"
,
...
...
vmdriver.py
View file @
2d1af23c
...
@@ -586,6 +586,6 @@ def get_node_metrics():
...
@@ -586,6 +586,6 @@ def get_node_metrics():
@celery.task
@celery.task
def
refresh_secret
(
user
,
secret
):
def
refresh_secret
(
user
):
ceph
.
delete_secret
(
user
)
ceph
.
delete_secret
(
user
)
ceph
.
check_secret
(
user
,
secret
)
ceph
.
check_secret
(
user
)
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment