root / snf-cyclades-app / synnefo / api / util.py @ a1713485
History | View | Annotate | Download (18.2 kB)
1 |
# Copyright 2011-2012 GRNET S.A. All rights reserved.
|
---|---|
2 |
#
|
3 |
# Redistribution and use in source and binary forms, with or
|
4 |
# without modification, are permitted provided that the following
|
5 |
# conditions are met:
|
6 |
#
|
7 |
# 1. Redistributions of source code must retain the above
|
8 |
# copyright notice, this list of conditions and the following
|
9 |
# disclaimer.
|
10 |
#
|
11 |
# 2. Redistributions in binary form must reproduce the above
|
12 |
# copyright notice, this list of conditions and the following
|
13 |
# disclaimer in the documentation and/or other materials
|
14 |
# provided with the distribution.
|
15 |
#
|
16 |
# THIS SOFTWARE IS PROVIDED BY GRNET S.A. ``AS IS'' AND ANY EXPRESS
|
17 |
# OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
18 |
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
19 |
# PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL GRNET S.A OR
|
20 |
# CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
21 |
# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
22 |
# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF
|
23 |
# USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
|
24 |
# AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
25 |
# LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
|
26 |
# ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
27 |
# POSSIBILITY OF SUCH DAMAGE.
|
28 |
#
|
29 |
# The views and conclusions contained in the software and
|
30 |
# documentation are those of the authors and should not be
|
31 |
# interpreted as representing official policies, either expressed
|
32 |
# or implied, of GRNET S.A.
|
33 |
|
34 |
from base64 import b64encode, b64decode |
35 |
from hashlib import sha256 |
36 |
from logging import getLogger |
37 |
from random import choice |
38 |
from string import digits, lowercase, uppercase |
39 |
|
40 |
from Crypto.Cipher import AES |
41 |
|
42 |
from django.conf import settings |
43 |
from django.http import HttpResponse |
44 |
from django.template.loader import render_to_string |
45 |
from django.utils import simplejson as json |
46 |
from django.db.models import Q |
47 |
|
48 |
from snf_django.lib.api import faults |
49 |
from synnefo.db.models import (Flavor, VirtualMachine, VirtualMachineMetadata, |
50 |
Network, NetworkInterface, SecurityGroup, |
51 |
BridgePoolTable, MacPrefixPoolTable, IPAddress, |
52 |
IPPoolTable) |
53 |
from synnefo.db import pools |
54 |
|
55 |
from synnefo.plankton.utils import image_backend |
56 |
|
57 |
from synnefo.cyclades_settings import cyclades_services, BASE_HOST |
58 |
from synnefo.lib.services import get_service_path |
59 |
from synnefo.lib import join_urls |
60 |
|
61 |
COMPUTE_URL = \ |
62 |
join_urls(BASE_HOST, |
63 |
get_service_path(cyclades_services, "compute", version="v2.0")) |
64 |
SERVERS_URL = join_urls(COMPUTE_URL, "servers/")
|
65 |
NETWORKS_URL = join_urls(COMPUTE_URL, "networks/")
|
66 |
FLAVORS_URL = join_urls(COMPUTE_URL, "flavors/")
|
67 |
IMAGES_URL = join_urls(COMPUTE_URL, "images/")
|
68 |
PLANKTON_URL = \ |
69 |
join_urls(BASE_HOST, |
70 |
get_service_path(cyclades_services, "image", version="v1.0")) |
71 |
IMAGES_PLANKTON_URL = join_urls(PLANKTON_URL, "images/")
|
72 |
|
73 |
PITHOSMAP_PREFIX = "pithosmap://"
|
74 |
|
75 |
log = getLogger('synnefo.api')
|
76 |
|
77 |
|
78 |
def random_password(): |
79 |
"""Generates a random password
|
80 |
|
81 |
We generate a windows compliant password: it must contain at least
|
82 |
one charachter from each of the groups: upper case, lower case, digits.
|
83 |
"""
|
84 |
|
85 |
pool = lowercase + uppercase + digits |
86 |
lowerset = set(lowercase)
|
87 |
upperset = set(uppercase)
|
88 |
digitset = set(digits)
|
89 |
length = 10
|
90 |
|
91 |
password = ''.join(choice(pool) for i in range(length - 2)) |
92 |
|
93 |
# Make sure the password is compliant
|
94 |
chars = set(password)
|
95 |
if not chars & lowerset: |
96 |
password += choice(lowercase) |
97 |
if not chars & upperset: |
98 |
password += choice(uppercase) |
99 |
if not chars & digitset: |
100 |
password += choice(digits) |
101 |
|
102 |
# Pad if necessary to reach required length
|
103 |
password += ''.join(choice(pool) for i in range(length - len(password))) |
104 |
|
105 |
return password
|
106 |
|
107 |
|
108 |
def zeropad(s): |
109 |
"""Add zeros at the end of a string in order to make its length
|
110 |
a multiple of 16."""
|
111 |
|
112 |
npad = 16 - len(s) % 16 |
113 |
return s + '\x00' * npad |
114 |
|
115 |
|
116 |
def encrypt(plaintext): |
117 |
# Make sure key is 32 bytes long
|
118 |
key = sha256(settings.SECRET_KEY).digest() |
119 |
|
120 |
aes = AES.new(key) |
121 |
enc = aes.encrypt(zeropad(plaintext)) |
122 |
return b64encode(enc)
|
123 |
|
124 |
|
125 |
def get_vm(server_id, user_id, for_update=False, non_deleted=False, |
126 |
non_suspended=False, prefetch_related=None): |
127 |
"""Find a VirtualMachine instance based on ID and owner."""
|
128 |
|
129 |
try:
|
130 |
server_id = int(server_id)
|
131 |
servers = VirtualMachine.objects |
132 |
if for_update:
|
133 |
servers = servers.select_for_update() |
134 |
if prefetch_related is not None: |
135 |
servers = servers.prefetch_related(prefetch_related) |
136 |
vm = servers.get(id=server_id, userid=user_id) |
137 |
if non_deleted and vm.deleted: |
138 |
raise faults.BadRequest("Server has been deleted.") |
139 |
if non_suspended and vm.suspended: |
140 |
raise faults.Forbidden("Administratively Suspended VM") |
141 |
return vm
|
142 |
except ValueError: |
143 |
raise faults.BadRequest('Invalid server ID.') |
144 |
except VirtualMachine.DoesNotExist:
|
145 |
raise faults.ItemNotFound('Server not found.') |
146 |
|
147 |
|
148 |
def get_vm_meta(vm, key): |
149 |
"""Return a VirtualMachineMetadata instance or raise ItemNotFound."""
|
150 |
|
151 |
try:
|
152 |
return VirtualMachineMetadata.objects.get(meta_key=key, vm=vm)
|
153 |
except VirtualMachineMetadata.DoesNotExist:
|
154 |
raise faults.ItemNotFound('Metadata key not found.') |
155 |
|
156 |
|
157 |
def get_image(image_id, user_id): |
158 |
"""Return an Image instance or raise ItemNotFound."""
|
159 |
|
160 |
with image_backend(user_id) as backend: |
161 |
return backend.get_image(image_id)
|
162 |
|
163 |
|
164 |
def get_image_dict(image_id, user_id): |
165 |
image = {} |
166 |
img = get_image(image_id, user_id) |
167 |
image["id"] = img["id"] |
168 |
image["name"] = img["name"] |
169 |
image["format"] = img["disk_format"] |
170 |
image["checksum"] = img["checksum"] |
171 |
image["location"] = img["location"] |
172 |
|
173 |
checksum = image["checksum"] = img["checksum"] |
174 |
size = image["size"] = img["size"] |
175 |
image["backend_id"] = PITHOSMAP_PREFIX + "/".join([checksum, str(size)]) |
176 |
|
177 |
properties = img.get("properties", {})
|
178 |
image["metadata"] = dict((key.upper(), val) |
179 |
for key, val in properties.items()) |
180 |
|
181 |
return image
|
182 |
|
183 |
|
184 |
def get_flavor(flavor_id, include_deleted=False): |
185 |
"""Return a Flavor instance or raise ItemNotFound."""
|
186 |
|
187 |
try:
|
188 |
flavor_id = int(flavor_id)
|
189 |
if include_deleted:
|
190 |
return Flavor.objects.get(id=flavor_id)
|
191 |
else:
|
192 |
return Flavor.objects.get(id=flavor_id, deleted=include_deleted)
|
193 |
except (ValueError, Flavor.DoesNotExist): |
194 |
raise faults.ItemNotFound('Flavor not found.') |
195 |
|
196 |
|
197 |
def get_flavor_provider(flavor): |
198 |
"""Extract provider from disk template.
|
199 |
|
200 |
Provider for `ext` disk_template is encoded in the disk template
|
201 |
name, which is formed `ext_<provider_name>`. Provider is None
|
202 |
for all other disk templates.
|
203 |
|
204 |
"""
|
205 |
disk_template = flavor.disk_template |
206 |
provider = None
|
207 |
if disk_template.startswith("ext"): |
208 |
disk_template, provider = disk_template.split("_", 1) |
209 |
return disk_template, provider
|
210 |
|
211 |
|
212 |
def get_network(network_id, user_id, for_update=False, non_deleted=False): |
213 |
"""Return a Network instance or raise ItemNotFound."""
|
214 |
|
215 |
try:
|
216 |
network_id = int(network_id)
|
217 |
objects = Network.objects.prefetch_related("subnets")
|
218 |
if for_update:
|
219 |
objects = objects.select_for_update() |
220 |
network = objects.get(Q(userid=user_id) | Q(public=True),
|
221 |
id=network_id) |
222 |
if non_deleted and network.deleted: |
223 |
raise faults.BadRequest("Network has been deleted.") |
224 |
return network
|
225 |
except (ValueError, Network.DoesNotExist): |
226 |
raise faults.ItemNotFound('Network %s not found.' % network_id) |
227 |
|
228 |
|
229 |
def get_port(port_id, user_id, for_update=False): |
230 |
"""
|
231 |
Return a NetworkInteface instance or raise ItemNotFound.
|
232 |
"""
|
233 |
try:
|
234 |
objects = NetworkInterface.objects |
235 |
if for_update:
|
236 |
objects = objects.select_for_update() |
237 |
|
238 |
port = objects.get(network__userid=user_id, id=port_id) |
239 |
|
240 |
if (port.device_owner != "vm") and for_update: |
241 |
raise faults.BadRequest('Can not update non vm port') |
242 |
|
243 |
return port
|
244 |
except (ValueError, NetworkInterface.DoesNotExist): |
245 |
raise faults.ItemNotFound('Port not found.') |
246 |
|
247 |
def get_security_group(sg_id): |
248 |
try:
|
249 |
sg = SecurityGroup.objects.get(id=sg_id) |
250 |
return sg
|
251 |
except (ValueError, SecurityGroup.DoesNotExist): |
252 |
raise faults.ItemNotFound("Not valid security group") |
253 |
|
254 |
def get_floating_ip_by_address(userid, address, for_update=False): |
255 |
try:
|
256 |
objects = IPAddress.objects |
257 |
if for_update:
|
258 |
objects = objects.select_for_update() |
259 |
return objects.get(userid=userid, floating_ip=True, |
260 |
address=address, deleted=False)
|
261 |
except IPAddress.DoesNotExist:
|
262 |
raise faults.ItemNotFound("Floating IP does not exist.") |
263 |
|
264 |
|
265 |
def get_floating_ip_by_id(userid, floating_ip_id, for_update=False): |
266 |
try:
|
267 |
objects = IPAddress.objects |
268 |
if for_update:
|
269 |
objects = objects.select_for_update() |
270 |
return objects.get(id=floating_ip_id, floating_ip=True, userid=userid, |
271 |
deleted=False)
|
272 |
except IPAddress.DoesNotExist:
|
273 |
raise faults.ItemNotFound("Floating IP %s does not exist." % |
274 |
floating_ip_id) |
275 |
|
276 |
|
277 |
def allocate_ip_from_pools(pool_rows, userid, address=None, floating_ip=False): |
278 |
"""Try to allocate a value from a number of pools.
|
279 |
|
280 |
This function takes as argument a number of PoolTable objects and tries to
|
281 |
allocate a value from them. If all pools are empty EmptyPool is raised.
|
282 |
|
283 |
"""
|
284 |
for pool_row in pool_rows: |
285 |
pool = pool_row.pool |
286 |
try:
|
287 |
value = pool.get(value=address) |
288 |
pool.save() |
289 |
subnet = pool_row.subnet |
290 |
ipaddress = IPAddress.objects.create(subnet=subnet, |
291 |
network=subnet.network, |
292 |
userid=userid, |
293 |
address=value, |
294 |
floating_ip=floating_ip) |
295 |
return ipaddress
|
296 |
except pools.EmptyPool:
|
297 |
pass
|
298 |
raise pools.EmptyPool("No more IP addresses available on pools %s" % |
299 |
pool_rows) |
300 |
|
301 |
|
302 |
def allocate_ip(network, userid, address=None, floating_ip=False): |
303 |
"""Try to allocate an IP from networks IP pools."""
|
304 |
ip_pools = IPPoolTable.objects.select_for_update()\ |
305 |
.filter(subnet__network=network) |
306 |
try:
|
307 |
return allocate_ip_from_pools(ip_pools, userid, address=address,
|
308 |
floating_ip=floating_ip) |
309 |
except pools.EmptyPool:
|
310 |
raise faults.Conflict("No more IP addresses available on network %s" |
311 |
% network.id) |
312 |
except pools.ValueNotAvailable:
|
313 |
raise faults.Conflict("IP address %s is already used." % address) |
314 |
except pools.InvalidValue:
|
315 |
raise faults.BadRequest("Address %s does not belong to network %s" % |
316 |
(address, network.id)) |
317 |
|
318 |
|
319 |
def allocate_public_ip(userid, floating_ip=False, backend=None): |
320 |
"""Try to allocate a public or floating IP address.
|
321 |
|
322 |
Try to allocate a a public IPv4 address from one of the available networks.
|
323 |
If 'floating_ip' is set, only networks which are floating IP pools will be
|
324 |
used and the IPAddress that will be created will be marked as a floating
|
325 |
IP. If 'backend' is set, only the networks that exist in this backend will
|
326 |
be used.
|
327 |
|
328 |
"""
|
329 |
|
330 |
ip_pool_rows = IPPoolTable.objects.select_for_update()\ |
331 |
.prefetch_related("subnet__network")\
|
332 |
.filter(subnet__deleted=False)\
|
333 |
.filter(subnet__network__public=True)\
|
334 |
.filter(subnet__network__drained=False)
|
335 |
if floating_ip:
|
336 |
ip_pool_rows = ip_pool_rows\ |
337 |
.filter(subnet__network__floating_ip_pool=True)
|
338 |
if backend is not None: |
339 |
ip_pool_rows = ip_pool_rows\ |
340 |
.filter(subnet__network__backend_networks__backend=backend) |
341 |
|
342 |
try:
|
343 |
return allocate_ip_from_pools(ip_pool_rows, userid,
|
344 |
floating_ip=floating_ip) |
345 |
except pools.EmptyPool:
|
346 |
ip_type = "floating" if floating_ip else "public" |
347 |
log_msg = "Failed to allocate a %s IP. Reason:" % ip_type
|
348 |
if ip_pool_rows:
|
349 |
log_msg += " No network exists."
|
350 |
else:
|
351 |
log_msg += " All network are full."
|
352 |
if backend is not None: |
353 |
log_msg += " Backend: %s" % backend
|
354 |
log.error(log_msg) |
355 |
exception_msg = "Can not allocate a %s IP address." % ip_type
|
356 |
raise faults.ServiceUnavailable(exception_msg)
|
357 |
|
358 |
|
359 |
def backend_has_free_public_ip(backend): |
360 |
"""Check if a backend has a free public IPv4 address."""
|
361 |
ip_pool_rows = IPPoolTable.objects.select_for_update()\ |
362 |
.filter(subnet__network__public=True)\
|
363 |
.filter(subnet__network__drained=False)\
|
364 |
.filter(subnet__deleted=False)\
|
365 |
.filter(subnet__network__backend_networks__backend=backend) |
366 |
for pool_row in ip_pool_rows: |
367 |
pool = pool_row.pool |
368 |
if pool.empty():
|
369 |
continue
|
370 |
else:
|
371 |
return True |
372 |
|
373 |
|
374 |
def backend_public_networks(backend): |
375 |
return Network.objects.filter(deleted=False, public=True, |
376 |
backend_networks__backend=backend) |
377 |
|
378 |
|
379 |
def get_vm_nic(vm, nic_id): |
380 |
"""Get a VMs NIC by its ID."""
|
381 |
try:
|
382 |
return vm.nics.get(id=nic_id)
|
383 |
except NetworkInterface.DoesNotExist:
|
384 |
raise faults.ItemNotFound("NIC '%s' not found" % nic_id) |
385 |
|
386 |
|
387 |
def get_nic(nic_id): |
388 |
try:
|
389 |
return NetworkInterface.objects.get(id=nic_id)
|
390 |
except NetworkInterface.DoesNotExist:
|
391 |
raise faults.ItemNotFound("NIC '%s' not found" % nic_id) |
392 |
|
393 |
|
394 |
def render_metadata(request, metadata, use_values=False, status=200): |
395 |
if request.serialization == 'xml': |
396 |
data = render_to_string('metadata.xml', {'metadata': metadata}) |
397 |
else:
|
398 |
if use_values:
|
399 |
d = {'metadata': {'values': metadata}} |
400 |
else:
|
401 |
d = {'metadata': metadata}
|
402 |
data = json.dumps(d) |
403 |
return HttpResponse(data, status=status)
|
404 |
|
405 |
|
406 |
def render_meta(request, meta, status=200): |
407 |
if request.serialization == 'xml': |
408 |
key, val = meta.items()[0]
|
409 |
data = render_to_string('meta.xml', dict(key=key, val=val)) |
410 |
else:
|
411 |
data = json.dumps(dict(meta=meta))
|
412 |
return HttpResponse(data, status=status)
|
413 |
|
414 |
|
415 |
def verify_personality(personality): |
416 |
"""Verify that a a list of personalities is well formed"""
|
417 |
if len(personality) > settings.MAX_PERSONALITY: |
418 |
raise faults.OverLimit("Maximum number of personalities" |
419 |
" exceeded")
|
420 |
for p in personality: |
421 |
# Verify that personalities are well-formed
|
422 |
try:
|
423 |
assert isinstance(p, dict) |
424 |
keys = set(p.keys())
|
425 |
allowed = set(['contents', 'group', 'mode', 'owner', 'path']) |
426 |
assert keys.issubset(allowed)
|
427 |
contents = p['contents']
|
428 |
if len(contents) > settings.MAX_PERSONALITY_SIZE: |
429 |
# No need to decode if contents already exceed limit
|
430 |
raise faults.OverLimit("Maximum size of personality exceeded") |
431 |
if len(b64decode(contents)) > settings.MAX_PERSONALITY_SIZE: |
432 |
raise faults.OverLimit("Maximum size of personality exceeded") |
433 |
except AssertionError: |
434 |
raise faults.BadRequest("Malformed personality in request") |
435 |
|
436 |
|
437 |
def values_from_flavor(flavor): |
438 |
"""Get Ganeti connectivity info from flavor type.
|
439 |
|
440 |
If link or mac_prefix equals to "pool", then the resources
|
441 |
are allocated from the corresponding Pools.
|
442 |
|
443 |
"""
|
444 |
try:
|
445 |
flavor = Network.FLAVORS[flavor] |
446 |
except KeyError: |
447 |
raise faults.BadRequest("Unknown network flavor") |
448 |
|
449 |
mode = flavor.get("mode")
|
450 |
|
451 |
link = flavor.get("link")
|
452 |
if link == "pool": |
453 |
link = allocate_resource("bridge")
|
454 |
|
455 |
mac_prefix = flavor.get("mac_prefix")
|
456 |
if mac_prefix == "pool": |
457 |
mac_prefix = allocate_resource("mac_prefix")
|
458 |
|
459 |
tags = flavor.get("tags")
|
460 |
|
461 |
return mode, link, mac_prefix, tags
|
462 |
|
463 |
|
464 |
def allocate_resource(res_type): |
465 |
table = get_pool_table(res_type) |
466 |
pool = table.get_pool() |
467 |
value = pool.get() |
468 |
pool.save() |
469 |
return value
|
470 |
|
471 |
|
472 |
def release_resource(res_type, value): |
473 |
table = get_pool_table(res_type) |
474 |
pool = table.get_pool() |
475 |
pool.put(value) |
476 |
pool.save() |
477 |
|
478 |
|
479 |
def get_pool_table(res_type): |
480 |
if res_type == "bridge": |
481 |
return BridgePoolTable
|
482 |
elif res_type == "mac_prefix": |
483 |
return MacPrefixPoolTable
|
484 |
else:
|
485 |
raise Exception("Unknown resource type") |
486 |
|
487 |
|
488 |
def get_existing_users(): |
489 |
"""
|
490 |
Retrieve user ids stored in cyclades user agnostic models.
|
491 |
"""
|
492 |
# also check PublicKeys a user with no servers/networks exist
|
493 |
from synnefo.userdata.models import PublicKeyPair |
494 |
from synnefo.db.models import VirtualMachine, Network |
495 |
|
496 |
keypairusernames = PublicKeyPair.objects.filter().values_list('user',
|
497 |
flat=True)
|
498 |
serverusernames = VirtualMachine.objects.filter().values_list('userid',
|
499 |
flat=True)
|
500 |
networkusernames = Network.objects.filter().values_list('userid',
|
501 |
flat=True)
|
502 |
|
503 |
return set(list(keypairusernames) + list(serverusernames) + |
504 |
list(networkusernames))
|
505 |
|
506 |
|
507 |
def vm_to_links(vm_id): |
508 |
href = join_urls(SERVERS_URL, str(vm_id))
|
509 |
return [{"rel": rel, "href": href} for rel in ("self", "bookmark")] |
510 |
|
511 |
|
512 |
def network_to_links(network_id): |
513 |
href = join_urls(NETWORKS_URL, str(network_id))
|
514 |
return [{"rel": rel, "href": href} for rel in ("self", "bookmark")] |
515 |
|
516 |
|
517 |
def flavor_to_links(flavor_id): |
518 |
href = join_urls(FLAVORS_URL, str(flavor_id))
|
519 |
return [{"rel": rel, "href": href} for rel in ("self", "bookmark")] |
520 |
|
521 |
|
522 |
def image_to_links(image_id): |
523 |
href = join_urls(IMAGES_URL, str(image_id))
|
524 |
links = [{"rel": rel, "href": href} for rel in ("self", "bookmark")] |
525 |
links.append({"rel": "alternate", |
526 |
"href": join_urls(IMAGES_PLANKTON_URL, str(image_id))}) |
527 |
return links
|
528 |
|
529 |
|
530 |
def start_action(vm, action, jobId): |
531 |
vm.action = action |
532 |
vm.backendjobid = jobId |
533 |
vm.backendopcode = None
|
534 |
vm.backendjobstatus = None
|
535 |
vm.backendlogmsg = None
|
536 |
vm.save() |