Files
watcher/watcher/tests/decision_engine/model/faker_cluster_state.py
licanwei 3bcd25727f update host_maintenance strategy
For Compute node, we can use the new property to calculate
resource(VCPU, memory and disk).

Partially Implements: blueprint improve-compute-data-model
Depends-on: I3f9a3279a26f3df444117d9265e74cca57b38d6e
Change-Id: I2bb230b5f5a573fb3045261dfdee73f1a8434e0d
2019-08-09 14:15:52 +08:00

359 lines
12 KiB
Python

# -*- encoding: utf-8 -*-
# Copyright (c) 2015 b<>com
#
# Authors: Jean-Emile DARTOIS <jean-emile.dartois@b-com.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import os
import mock
from watcher.common import utils
from watcher.decision_engine.model.collector import base
from watcher.decision_engine.model import element
from watcher.decision_engine.model import model_root as modelroot
volume_uuid_mapping = {
"volume_0": "5028b1eb-8749-48ae-a42c-5bdd1323976f",
"volume_1": "74454247-a064-4b34-8f43-89337987720e",
"volume_2": "a16c811e-2521-4fd3-8779-6a94ccb3be73",
"volume_3": "37856b95-5be4-4864-8a49-c83f55c66780",
}
class FakerModelCollector(base.BaseClusterDataModelCollector):
def __init__(self, config=None, osc=None, audit_scope=None):
if config is None:
config = mock.Mock(period=777)
super(FakerModelCollector, self).__init__(config)
@property
def notification_endpoints(self):
return []
def load_data(self, filename):
cwd = os.path.abspath(os.path.dirname(__file__))
data_folder = os.path.join(cwd, "data")
with open(os.path.join(data_folder, filename), 'rb') as xml_file:
xml_data = xml_file.read()
return xml_data
def load_model(self, filename):
return modelroot.ModelRoot.from_xml(self.load_data(filename))
def get_audit_scope_handler(self, audit_scope):
return None
def execute(self):
return self._cluster_data_model or self.build_scenario_1()
def build_scenario_1(self):
instances = []
model = modelroot.ModelRoot()
# number of nodes
node_count = 5
# number max of instance per node
node_instance_count = 7
# total number of virtual machine
instance_count = (node_count * node_instance_count)
for id_ in range(0, node_count):
node_uuid = "Node_{0}".format(id_)
hostname = "hostname_{0}".format(id_)
node_attributes = {
"id": id_,
"uuid": node_uuid,
"hostname": hostname,
"memory": 132,
"memory_mb_reserved": 0,
"memory_ratio": 1,
"disk": 250,
"disk_capacity": 250,
"disk_gb_reserved": 0,
"disk_ratio": 1,
"vcpus": 40,
"vcpu_reserved": 0,
"vcpu_ratio": 1,
}
node = element.ComputeNode(**node_attributes)
model.add_node(node)
for i in range(0, instance_count):
instance_uuid = "INSTANCE_{0}".format(i)
if instance_uuid == "INSTANCE_1":
project_id = "26F03131-32CB-4697-9D61-9123F87A8147"
elif instance_uuid == "INSTANCE_2":
project_id = "109F7909-0607-4712-B32C-5CC6D49D2F15"
else:
project_id = "91FFFE30-78A0-4152-ACD2-8310FF274DC9"
instance_attributes = {
"uuid": instance_uuid,
"name": instance_uuid,
"memory": 2,
"disk": 20,
"disk_capacity": 20,
"vcpus": 10,
"metadata":
'{"optimize": true,"top": "floor","nested": {"x": "y"}}',
"project_id": project_id
}
instance = element.Instance(**instance_attributes)
instances.append(instance)
model.add_instance(instance)
mappings = [
("INSTANCE_0", "Node_0"),
("INSTANCE_1", "Node_0"),
("INSTANCE_2", "Node_1"),
("INSTANCE_3", "Node_2"),
("INSTANCE_4", "Node_2"),
("INSTANCE_5", "Node_2"),
("INSTANCE_6", "Node_3"),
("INSTANCE_7", "Node_4"),
]
for instance_uuid, node_uuid in mappings:
model.map_instance(
model.get_instance_by_uuid(instance_uuid),
model.get_node_by_uuid(node_uuid),
)
return model
def generate_scenario_1(self):
return self.load_model('scenario_1.xml')
def generate_scenario_1_with_1_node_unavailable(self):
return self.load_model('scenario_1_with_1_node_unavailable.xml')
def generate_scenario_1_with_all_nodes_disable(self):
return self.load_model('scenario_1_with_all_nodes_disable.xml')
def generate_scenario_1_with_all_instances_exclude(self):
return self.load_model('scenario_1_with_all_instances_exclude.xml')
def generate_scenario_3_with_2_nodes(self):
return self.load_model('scenario_3_with_2_nodes.xml')
def generate_scenario_4_with_1_node_no_instance(self):
return self.load_model('scenario_4_with_1_node_no_instance.xml')
def generate_scenario_5_with_instance_disk_0(self):
return self.load_model('scenario_5_with_instance_disk_0.xml')
def generate_scenario_6_with_2_nodes(self):
return self.load_model('scenario_6_with_2_nodes.xml')
def generate_scenario_7_with_2_nodes(self):
return self.load_model('scenario_7_with_2_nodes.xml')
def generate_scenario_8_with_4_nodes(self):
return self.load_model('scenario_8_with_4_nodes.xml')
def generate_scenario_9_with_3_active_plus_1_disabled_nodes(self):
return self.load_model(
'scenario_9_with_3_active_plus_1_disabled_nodes.xml')
class FakerStorageModelCollector(base.BaseClusterDataModelCollector):
def __init__(self, config=None, osc=None, audit_scope=None):
if config is None:
config = mock.Mock(period=777)
super(FakerStorageModelCollector, self).__init__(config)
@property
def notification_endpoints(self):
return []
def load_data(self, filename):
cwd = os.path.abspath(os.path.dirname(__file__))
data_folder = os.path.join(cwd, "data")
with open(os.path.join(data_folder, filename), 'rb') as xml_file:
xml_data = xml_file.read()
return xml_data
def load_model(self, filename):
return modelroot.StorageModelRoot.from_xml(self.load_data(filename))
def get_audit_scope_handler(self, audit_scope):
return None
def execute(self):
return self._cluster_data_model or self.build_scenario_1()
def build_scenario_1(self):
model = modelroot.StorageModelRoot()
# number of nodes
node_count = 2
# number of pools per node
pool_count = 2
# number of volumes
volume_count = 9
for i in range(0, node_count):
host = "host_{0}@backend_{0}".format(i)
zone = "zone_{0}".format(i)
volume_type = ["type_{0}".format(i)]
node_attributes = {
"host": host,
"zone": zone,
"status": 'enabled',
"state": 'up',
"volume_type": volume_type,
}
node = element.StorageNode(**node_attributes)
model.add_node(node)
for j in range(0, pool_count):
name = "host_{0}@backend_{0}#pool_{1}".format(i, j)
pool_attributes = {
"name": name,
"total_volumes": 2,
"total_capacity_gb": 500,
"free_capacity_gb": 420,
"provisioned_capacity_gb": 80,
"allocated_capacity_gb": 80,
"virtual_free": 420,
}
pool = element.Pool(**pool_attributes)
model.add_pool(pool)
mappings = [
("host_0@backend_0#pool_0", "host_0@backend_0"),
("host_0@backend_0#pool_1", "host_0@backend_0"),
("host_1@backend_1#pool_0", "host_1@backend_1"),
("host_1@backend_1#pool_1", "host_1@backend_1"),
]
for pool_name, node_name in mappings:
model.map_pool(
model.get_pool_by_pool_name(pool_name),
model.get_node_by_name(node_name),
)
volume_uuid_mapping = [
"5028b1eb-8749-48ae-a42c-5bdd1323976f",
"74454247-a064-4b34-8f43-89337987720e",
"a16c811e-2521-4fd3-8779-6a94ccb3be73",
"37856b95-5be4-4864-8a49-c83f55c66780",
"694f8fb1-df96-46be-b67d-49f2c14a495e",
"66b094b0-8fc3-4a94-913f-a5f9312b11a5",
"e9013810-4b4c-4b94-a056-4c36702d51a3",
"07976191-6a57-4c35-9f3c-55b3b5ecd6d5",
"4d1c952d-95d0-4aac-82aa-c3cb509af9f3",
]
for k in range(volume_count):
uuid = volume_uuid_mapping[k]
name = "name_{0}".format(k)
volume_attributes = {
"size": 40,
"status": "in-use",
"uuid": uuid,
"attachments":
'[{"server_id": "server","attachment_id": "attachment"}]',
"name": name,
"multiattach": 'True',
"snapshot_id": uuid,
"project_id": "91FFFE30-78A0-4152-ACD2-8310FF274DC9",
"metadata": '{"readonly": false,"attached_mode": "rw"}',
"bootable": 'False'
}
volume = element.Volume(**volume_attributes)
model.add_volume(volume)
mappings = [
(volume_uuid_mapping[0], "host_0@backend_0#pool_0"),
(volume_uuid_mapping[1], "host_0@backend_0#pool_0"),
(volume_uuid_mapping[2], "host_0@backend_0#pool_1"),
(volume_uuid_mapping[3], "host_0@backend_0#pool_1"),
(volume_uuid_mapping[4], "host_1@backend_1#pool_0"),
(volume_uuid_mapping[5], "host_1@backend_1#pool_0"),
(volume_uuid_mapping[6], "host_1@backend_1#pool_1"),
(volume_uuid_mapping[7], "host_1@backend_1#pool_1"),
]
for volume_uuid, pool_name in mappings:
model.map_volume(
model.get_volume_by_uuid(volume_uuid),
model.get_pool_by_pool_name(pool_name),
)
return model
def generate_scenario_1(self):
return self.load_model('storage_scenario_1.xml')
class FakerBaremetalModelCollector(base.BaseClusterDataModelCollector):
def __init__(self, config=None, osc=None):
if config is None:
config = mock.Mock(period=777)
super(FakerBaremetalModelCollector, self).__init__(config)
@property
def notification_endpoints(self):
return []
def get_audit_scope_handler(self, audit_scope):
return None
def load_data(self, filename):
cwd = os.path.abspath(os.path.dirname(__file__))
data_folder = os.path.join(cwd, "data")
with open(os.path.join(data_folder, filename), 'rb') as xml_file:
xml_data = xml_file.read()
return xml_data
def load_model(self, filename):
return modelroot.BaremetalModelRoot.from_xml(self.load_data(filename))
def execute(self):
return self._cluster_data_model or self.build_scenario_1()
def build_scenario_1(self):
model = modelroot.BaremetalModelRoot()
# number of nodes
node_count = 2
for i in range(0, node_count):
uuid = utils.generate_uuid()
node_attributes = {
"uuid": uuid,
"power_state": "power on",
"maintenance": "false",
"maintenance_reason": "null",
"extra": {"compute_node_id": i}
}
node = element.IronicNode(**node_attributes)
model.add_node(node)
return model
def generate_scenario_1(self):
return self.load_model('ironic_scenario_1.xml')