Compare commits
134 Commits
4.0.0.0rc1
...
12.0.0.0rc
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
6b433b3547 | ||
|
|
9d58a6d457 | ||
|
|
c95ce4ec17 | ||
|
|
9492c2190e | ||
|
|
808f1bcee3 | ||
|
|
3b224b5629 | ||
|
|
424e9a76af | ||
|
|
40e93407c7 | ||
|
|
721aec1cb6 | ||
|
|
8a3ee8f931 | ||
|
|
00fea975e2 | ||
|
|
fd6562382e | ||
|
|
ec90891636 | ||
|
|
7336a48057 | ||
|
|
922478fbda | ||
|
|
9f0eca2343 | ||
|
|
1e11c490a7 | ||
|
|
8a7a8db661 | ||
|
|
0610070e59 | ||
|
|
a0997a0423 | ||
|
|
4ea3eada3e | ||
|
|
cd1c0f3054 | ||
|
|
684350977d | ||
|
|
d28630b759 | ||
|
|
f7fbaf46a2 | ||
|
|
e7cda537e7 | ||
|
|
c7be34fbaa | ||
|
|
52da088011 | ||
|
|
6ac3a6febf | ||
|
|
e36b77ad6d | ||
|
|
6003322711 | ||
|
|
f4ffca01b8 | ||
|
|
5d70c207cd | ||
|
|
0b2e641d00 | ||
|
|
ff84b052a5 | ||
|
|
a43b040ebc | ||
|
|
749fa2507a | ||
|
|
76d61362ee | ||
|
|
c55143bc21 | ||
|
|
7609df3370 | ||
|
|
b57eac12cb | ||
|
|
ac6911d3c4 | ||
|
|
23c2010681 | ||
|
|
01d74d0a87 | ||
|
|
e4fab0ce7f | ||
|
|
76ecaaeb3a | ||
|
|
6dd2f2a9c1 | ||
|
|
a993849928 | ||
|
|
6dbac1f6ae | ||
|
|
c28756c48b | ||
|
|
2414f66e38 | ||
|
|
546b730c9b | ||
|
|
75be54aa89 | ||
|
|
4e9e75f4a0 | ||
|
|
e0779175cf | ||
|
|
1235153b4c | ||
|
|
bf5ce9aa3f | ||
|
|
204992ff6f | ||
|
|
386e288543 | ||
|
|
6f668133ad | ||
|
|
fb625bfa56 | ||
|
|
32cb132712 | ||
|
|
9ca44fa3ab | ||
|
|
2205f4e4e3 | ||
|
|
7cd4373707 | ||
|
|
a2123088bf | ||
|
|
863815153e | ||
|
|
76270c8383 | ||
|
|
58de9c405a | ||
|
|
8f0126f1fe | ||
|
|
ec21898978 | ||
|
|
e61f9b5e88 | ||
|
|
e91efbde01 | ||
|
|
63b6997c83 | ||
|
|
262edc8cc9 | ||
|
|
204b276693 | ||
|
|
f8a2877f24 | ||
|
|
af02bebca9 | ||
|
|
3aaa20908d | ||
|
|
5097665be3 | ||
|
|
09f6e3bde5 | ||
|
|
f488636fb8 | ||
|
|
11cb88c2cd | ||
|
|
16a0486655 | ||
|
|
2454d4d199 | ||
|
|
45dca00dee | ||
|
|
09b2383685 | ||
|
|
f8797a7f70 | ||
|
|
da283b49b8 | ||
|
|
e21e5f609e | ||
|
|
583c946061 | ||
|
|
cca0d9f7d7 | ||
|
|
25a0b184a1 | ||
|
|
ed59145354 | ||
|
|
19adfda3b9 | ||
|
|
fa56bc715e | ||
|
|
350ce66d3c | ||
|
|
1667046f58 | ||
|
|
3f7a508a2e | ||
|
|
f7f5659bca | ||
|
|
57f55190ff | ||
|
|
237550ad57 | ||
|
|
cad67702d6 | ||
|
|
ae678dfaaa | ||
|
|
5ad3960286 | ||
|
|
dbd86be363 | ||
|
|
9f0138e1cf | ||
|
|
097ac06f0b | ||
|
|
0869b1c75c | ||
|
|
527578a147 | ||
|
|
b0c411b22a | ||
|
|
4a1915bec4 | ||
|
|
751027858b | ||
|
|
12bd9c0590 | ||
|
|
1ff940598f | ||
|
|
9d495618d2 | ||
|
|
c6d2690aa3 | ||
|
|
623e44ecf9 | ||
|
|
5c34b6bc47 | ||
|
|
8a36ad5f87 | ||
|
|
6ff95efaf6 | ||
|
|
ba2f1804b0 | ||
|
|
44061326e9 | ||
|
|
0b4c4f1de6 | ||
|
|
9652571437 | ||
|
|
f0f15f89c6 | ||
|
|
075e374b3d | ||
|
|
eaa0dfea4b | ||
|
|
b7956de761 | ||
|
|
a30dbdd724 | ||
|
|
60a829e982 | ||
|
|
74cfa0fc8c | ||
|
|
5071c8f8fa | ||
|
|
0ef0f165cb |
48
.zuul.yaml
48
.zuul.yaml
@@ -1,9 +1,9 @@
|
|||||||
- project:
|
- project:
|
||||||
|
queue: watcher
|
||||||
templates:
|
templates:
|
||||||
- check-requirements
|
- check-requirements
|
||||||
- openstack-cover-jobs
|
- openstack-cover-jobs
|
||||||
- openstack-lower-constraints-jobs
|
- openstack-python3-jobs
|
||||||
- openstack-python3-ussuri-jobs
|
|
||||||
- publish-openstack-docs-pti
|
- publish-openstack-docs-pti
|
||||||
- release-notes-jobs-python3
|
- release-notes-jobs-python3
|
||||||
check:
|
check:
|
||||||
@@ -13,10 +13,8 @@
|
|||||||
- watcher-tempest-strategies
|
- watcher-tempest-strategies
|
||||||
- watcher-tempest-actuator
|
- watcher-tempest-actuator
|
||||||
- watcherclient-tempest-functional
|
- watcherclient-tempest-functional
|
||||||
- watcher-tls-test
|
|
||||||
- watcher-tempest-functional-ipv6-only
|
- watcher-tempest-functional-ipv6-only
|
||||||
gate:
|
gate:
|
||||||
queue: watcher
|
|
||||||
jobs:
|
jobs:
|
||||||
- watcher-tempest-functional
|
- watcher-tempest-functional
|
||||||
- watcher-tempest-functional-ipv6-only
|
- watcher-tempest-functional-ipv6-only
|
||||||
@@ -88,21 +86,10 @@
|
|||||||
tempest_concurrency: 1
|
tempest_concurrency: 1
|
||||||
tempest_test_regex: watcher_tempest_plugin.tests.scenario.test_execute_strategies
|
tempest_test_regex: watcher_tempest_plugin.tests.scenario.test_execute_strategies
|
||||||
|
|
||||||
- job:
|
|
||||||
name: watcher-tls-test
|
|
||||||
parent: watcher-tempest-multinode
|
|
||||||
group-vars:
|
|
||||||
subnode:
|
|
||||||
devstack_services:
|
|
||||||
tls-proxy: true
|
|
||||||
vars:
|
|
||||||
devstack_services:
|
|
||||||
tls-proxy: true
|
|
||||||
|
|
||||||
- job:
|
- job:
|
||||||
name: watcher-tempest-multinode
|
name: watcher-tempest-multinode
|
||||||
parent: watcher-tempest-functional
|
parent: watcher-tempest-functional
|
||||||
nodeset: openstack-two-node-bionic
|
nodeset: openstack-two-node-jammy
|
||||||
roles:
|
roles:
|
||||||
- zuul: openstack/tempest
|
- zuul: openstack/tempest
|
||||||
group-vars:
|
group-vars:
|
||||||
@@ -120,8 +107,7 @@
|
|||||||
watcher-api: false
|
watcher-api: false
|
||||||
watcher-decision-engine: true
|
watcher-decision-engine: true
|
||||||
watcher-applier: false
|
watcher-applier: false
|
||||||
# We need to add TLS support for watcher plugin
|
c-bak: false
|
||||||
tls-proxy: false
|
|
||||||
ceilometer: false
|
ceilometer: false
|
||||||
ceilometer-acompute: false
|
ceilometer-acompute: false
|
||||||
ceilometer-acentral: false
|
ceilometer-acentral: false
|
||||||
@@ -160,7 +146,6 @@
|
|||||||
timeout: 7200
|
timeout: 7200
|
||||||
required-projects: &base_required_projects
|
required-projects: &base_required_projects
|
||||||
- openstack/ceilometer
|
- openstack/ceilometer
|
||||||
- openstack/devstack-gate
|
|
||||||
- openstack/python-openstackclient
|
- openstack/python-openstackclient
|
||||||
- openstack/python-watcherclient
|
- openstack/python-watcherclient
|
||||||
- openstack/watcher
|
- openstack/watcher
|
||||||
@@ -170,7 +155,6 @@
|
|||||||
devstack_plugins:
|
devstack_plugins:
|
||||||
watcher: https://opendev.org/openstack/watcher
|
watcher: https://opendev.org/openstack/watcher
|
||||||
devstack_services:
|
devstack_services:
|
||||||
tls-proxy: false
|
|
||||||
watcher-api: true
|
watcher-api: true
|
||||||
watcher-decision-engine: true
|
watcher-decision-engine: true
|
||||||
watcher-applier: true
|
watcher-applier: true
|
||||||
@@ -179,14 +163,10 @@
|
|||||||
s-container: false
|
s-container: false
|
||||||
s-object: false
|
s-object: false
|
||||||
s-proxy: false
|
s-proxy: false
|
||||||
devstack_localrc:
|
tempest_plugins:
|
||||||
TEMPEST_PLUGINS: /opt/stack/watcher-tempest-plugin
|
- watcher-tempest-plugin
|
||||||
USE_PYTHON3: true
|
|
||||||
tempest_test_regex: watcher_tempest_plugin.tests.api
|
tempest_test_regex: watcher_tempest_plugin.tests.api
|
||||||
tox_envlist: all
|
tox_envlist: all
|
||||||
tox_environment:
|
|
||||||
# Do we really need to set this? It's cargo culted
|
|
||||||
PYTHONUNBUFFERED: 'true'
|
|
||||||
zuul_copy_output:
|
zuul_copy_output:
|
||||||
/etc/hosts: logs
|
/etc/hosts: logs
|
||||||
|
|
||||||
@@ -200,10 +180,12 @@
|
|||||||
|
|
||||||
- job:
|
- job:
|
||||||
name: watcher-grenade
|
name: watcher-grenade
|
||||||
parent: legacy-dsvm-base
|
parent: grenade
|
||||||
timeout: 10800
|
required-projects:
|
||||||
run: playbooks/legacy/grenade-devstack-watcher/run.yaml
|
- openstack/watcher
|
||||||
post-run: playbooks/legacy/grenade-devstack-watcher/post.yaml
|
- openstack/python-watcherclient
|
||||||
|
- openstack/watcher-tempest-plugin
|
||||||
|
vars: *base_vars
|
||||||
irrelevant-files:
|
irrelevant-files:
|
||||||
- ^(test-|)requirements.txt$
|
- ^(test-|)requirements.txt$
|
||||||
- ^.*\.rst$
|
- ^.*\.rst$
|
||||||
@@ -215,12 +197,6 @@
|
|||||||
- ^setup.cfg$
|
- ^setup.cfg$
|
||||||
- ^tools/.*$
|
- ^tools/.*$
|
||||||
- ^tox.ini$
|
- ^tox.ini$
|
||||||
required-projects:
|
|
||||||
- openstack/grenade
|
|
||||||
- openstack/devstack-gate
|
|
||||||
- openstack/watcher
|
|
||||||
- openstack/python-watcherclient
|
|
||||||
- openstack/watcher-tempest-plugin
|
|
||||||
|
|
||||||
- job:
|
- job:
|
||||||
# This job is used in python-watcherclient repo
|
# This job is used in python-watcherclient repo
|
||||||
|
|||||||
@@ -22,9 +22,6 @@
|
|||||||
# All configuration values have a default; values that are commented out
|
# All configuration values have a default; values that are commented out
|
||||||
# serve to show the default.
|
# serve to show the default.
|
||||||
|
|
||||||
from watcher import version as watcher_version
|
|
||||||
|
|
||||||
|
|
||||||
extensions = [
|
extensions = [
|
||||||
'openstackdocstheme',
|
'openstackdocstheme',
|
||||||
'os_api_ref',
|
'os_api_ref',
|
||||||
@@ -46,21 +43,13 @@ project = u'Infrastructure Optimization API Reference'
|
|||||||
copyright = u'2010-present, OpenStack Foundation'
|
copyright = u'2010-present, OpenStack Foundation'
|
||||||
|
|
||||||
# openstackdocstheme options
|
# openstackdocstheme options
|
||||||
repository_name = 'openstack/watcher'
|
openstackdocs_repo_name = 'openstack/watcher'
|
||||||
bug_project = 'watcher'
|
openstackdocs_auto_name = False
|
||||||
bug_tag = ''
|
openstackdocs_bug_project = 'watcher'
|
||||||
|
openstackdocs_bug_tag = ''
|
||||||
# The version info for the project you're documenting, acts as replacement for
|
|
||||||
# |version| and |release|, also used in various other places throughout the
|
|
||||||
# built documents.
|
|
||||||
#
|
|
||||||
# The full version, including alpha/beta/rc tags.
|
|
||||||
release = watcher_version.version_info.release_string()
|
|
||||||
# The short X.Y version.
|
|
||||||
version = watcher_version.version_string
|
|
||||||
|
|
||||||
# The name of the Pygments (syntax highlighting) style to use.
|
# The name of the Pygments (syntax highlighting) style to use.
|
||||||
pygments_style = 'sphinx'
|
pygments_style = 'native'
|
||||||
|
|
||||||
# -- Options for HTML output --------------------------------------------------
|
# -- Options for HTML output --------------------------------------------------
|
||||||
|
|
||||||
@@ -75,10 +64,6 @@ html_theme_options = {
|
|||||||
"sidebar_mode": "toc",
|
"sidebar_mode": "toc",
|
||||||
}
|
}
|
||||||
|
|
||||||
# If not '', a 'Last updated on:' timestamp is inserted at every page bottom,
|
|
||||||
# using the given strftime format.
|
|
||||||
html_last_updated_fmt = '%Y-%m-%d %H:%M'
|
|
||||||
|
|
||||||
# -- Options for LaTeX output -------------------------------------------------
|
# -- Options for LaTeX output -------------------------------------------------
|
||||||
|
|
||||||
# Grouping the document tree into LaTeX files. List of tuples
|
# Grouping the document tree into LaTeX files. List of tuples
|
||||||
|
|||||||
@@ -298,7 +298,7 @@ function start_watcher_api {
|
|||||||
service_protocol="http"
|
service_protocol="http"
|
||||||
fi
|
fi
|
||||||
if [[ "$WATCHER_USE_WSGI_MODE" == "uwsgi" ]]; then
|
if [[ "$WATCHER_USE_WSGI_MODE" == "uwsgi" ]]; then
|
||||||
run_process "watcher-api" "$WATCHER_BIN_DIR/uwsgi --procname-prefix watcher-api --ini $WATCHER_UWSGI_CONF"
|
run_process "watcher-api" "$(which uwsgi) --procname-prefix watcher-api --ini $WATCHER_UWSGI_CONF"
|
||||||
watcher_url=$service_protocol://$SERVICE_HOST/infra-optim
|
watcher_url=$service_protocol://$SERVICE_HOST/infra-optim
|
||||||
else
|
else
|
||||||
watcher_url=$service_protocol://$SERVICE_HOST:$service_port
|
watcher_url=$service_protocol://$SERVICE_HOST:$service_port
|
||||||
|
|||||||
@@ -13,8 +13,6 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
from __future__ import unicode_literals
|
|
||||||
|
|
||||||
import importlib
|
import importlib
|
||||||
import inspect
|
import inspect
|
||||||
|
|
||||||
|
|||||||
@@ -1,10 +1,10 @@
|
|||||||
# The order of packages is significant, because pip processes them in the order
|
# The order of packages is significant, because pip processes them in the order
|
||||||
# of appearance. Changing the order has an impact on the overall integration
|
# of appearance. Changing the order has an impact on the overall integration
|
||||||
# process, which may cause wedges in the gate later.
|
# process, which may cause wedges in the gate later.
|
||||||
openstackdocstheme>=1.20.0 # Apache-2.0
|
openstackdocstheme>=2.2.1 # Apache-2.0
|
||||||
sphinx>=1.8.0,!=2.1.0,!=3.0.0 # BSD
|
sphinx>=2.0.0,!=2.1.0 # BSD
|
||||||
sphinxcontrib-pecanwsme>=0.8.0 # Apache-2.0
|
sphinxcontrib-pecanwsme>=0.8.0 # Apache-2.0
|
||||||
sphinxcontrib-svg2pdfconverter>=0.1.0 # BSD
|
sphinxcontrib-svg2pdfconverter>=0.1.0 # BSD
|
||||||
reno>=2.7.0 # Apache-2.0
|
reno>=3.1.0 # Apache-2.0
|
||||||
sphinxcontrib-apidoc>=0.2.0 # BSD
|
sphinxcontrib-apidoc>=0.2.0 # BSD
|
||||||
os-api-ref>=1.4.0 # Apache-2.0
|
os-api-ref>=1.4.0 # Apache-2.0
|
||||||
|
|||||||
@@ -17,6 +17,14 @@
|
|||||||
Policies
|
Policies
|
||||||
========
|
========
|
||||||
|
|
||||||
|
.. warning::
|
||||||
|
|
||||||
|
JSON formatted policy file is deprecated since Watcher 6.0.0 (Wallaby).
|
||||||
|
This `oslopolicy-convert-json-to-yaml`__ tool will migrate your existing
|
||||||
|
JSON-formatted policy file to YAML in a backward-compatible way.
|
||||||
|
|
||||||
|
.. __: https://docs.openstack.org/oslo.policy/latest/cli/oslopolicy-convert-json-to-yaml.html
|
||||||
|
|
||||||
Watcher's public API calls may be restricted to certain sets of users using a
|
Watcher's public API calls may be restricted to certain sets of users using a
|
||||||
policy configuration file. This document explains exactly how policies are
|
policy configuration file. This document explains exactly how policies are
|
||||||
configured and what they apply to.
|
configured and what they apply to.
|
||||||
|
|||||||
@@ -14,7 +14,6 @@
|
|||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
from watcher import version as watcher_version
|
|
||||||
from watcher import objects
|
from watcher import objects
|
||||||
|
|
||||||
objects.register_all()
|
objects.register_all()
|
||||||
@@ -57,18 +56,8 @@ source_suffix = '.rst'
|
|||||||
master_doc = 'index'
|
master_doc = 'index'
|
||||||
|
|
||||||
# General information about the project.
|
# General information about the project.
|
||||||
project = u'Watcher'
|
project = 'Watcher'
|
||||||
copyright = u'OpenStack Foundation'
|
copyright = 'OpenStack Foundation'
|
||||||
|
|
||||||
# The version info for the project you're documenting, acts as replacement for
|
|
||||||
# |version| and |release|, also used in various other places throughout the
|
|
||||||
# built documents.
|
|
||||||
#
|
|
||||||
# The short X.Y version.
|
|
||||||
# The full version, including alpha/beta/rc tags.
|
|
||||||
release = watcher_version.version_info.release_string()
|
|
||||||
# The short X.Y version.
|
|
||||||
version = watcher_version.version_string
|
|
||||||
|
|
||||||
# A list of ignored prefixes for module index sorting.
|
# A list of ignored prefixes for module index sorting.
|
||||||
modindex_common_prefix = ['watcher.']
|
modindex_common_prefix = ['watcher.']
|
||||||
@@ -94,7 +83,7 @@ add_module_names = True
|
|||||||
suppress_warnings = ['app.add_directive']
|
suppress_warnings = ['app.add_directive']
|
||||||
|
|
||||||
# The name of the Pygments (syntax highlighting) style to use.
|
# The name of the Pygments (syntax highlighting) style to use.
|
||||||
pygments_style = 'sphinx'
|
pygments_style = 'native'
|
||||||
|
|
||||||
# -- Options for man page output --------------------------------------------
|
# -- Options for man page output --------------------------------------------
|
||||||
|
|
||||||
@@ -102,14 +91,14 @@ pygments_style = 'sphinx'
|
|||||||
# List of tuples 'sourcefile', 'target', u'title', u'Authors name', 'manual'
|
# List of tuples 'sourcefile', 'target', u'title', u'Authors name', 'manual'
|
||||||
|
|
||||||
man_pages = [
|
man_pages = [
|
||||||
('man/watcher-api', 'watcher-api', u'Watcher API Server',
|
('man/watcher-api', 'watcher-api', 'Watcher API Server',
|
||||||
[u'OpenStack'], 1),
|
['OpenStack'], 1),
|
||||||
('man/watcher-applier', 'watcher-applier', u'Watcher Applier',
|
('man/watcher-applier', 'watcher-applier', 'Watcher Applier',
|
||||||
[u'OpenStack'], 1),
|
['OpenStack'], 1),
|
||||||
('man/watcher-db-manage', 'watcher-db-manage',
|
('man/watcher-db-manage', 'watcher-db-manage',
|
||||||
u'Watcher Db Management Utility', [u'OpenStack'], 1),
|
'Watcher Db Management Utility', ['OpenStack'], 1),
|
||||||
('man/watcher-decision-engine', 'watcher-decision-engine',
|
('man/watcher-decision-engine', 'watcher-decision-engine',
|
||||||
u'Watcher Decision Engine', [u'OpenStack'], 1),
|
'Watcher Decision Engine', ['OpenStack'], 1),
|
||||||
]
|
]
|
||||||
|
|
||||||
# -- Options for HTML output --------------------------------------------------
|
# -- Options for HTML output --------------------------------------------------
|
||||||
@@ -125,12 +114,13 @@ html_theme = 'openstackdocs'
|
|||||||
# Output file base name for HTML help builder.
|
# Output file base name for HTML help builder.
|
||||||
htmlhelp_basename = '%sdoc' % project
|
htmlhelp_basename = '%sdoc' % project
|
||||||
|
|
||||||
html_last_updated_fmt = '%Y-%m-%d %H:%M'
|
|
||||||
|
|
||||||
#openstackdocstheme options
|
#openstackdocstheme options
|
||||||
repository_name = 'openstack/watcher'
|
openstackdocs_repo_name = 'openstack/watcher'
|
||||||
bug_project = 'watcher'
|
openstackdocs_pdf_link = True
|
||||||
bug_tag = ''
|
openstackdocs_auto_name = False
|
||||||
|
openstackdocs_bug_project = 'watcher'
|
||||||
|
openstackdocs_bug_tag = ''
|
||||||
|
|
||||||
# Grouping the document tree into LaTeX files. List of tuples
|
# Grouping the document tree into LaTeX files. List of tuples
|
||||||
# (source start file, target name, title, author, documentclass
|
# (source start file, target name, title, author, documentclass
|
||||||
@@ -138,8 +128,8 @@ bug_tag = ''
|
|||||||
latex_documents = [
|
latex_documents = [
|
||||||
('index',
|
('index',
|
||||||
'doc-watcher.tex',
|
'doc-watcher.tex',
|
||||||
u'%s Documentation' % project,
|
'Watcher Documentation',
|
||||||
u'OpenStack Foundation', 'manual'),
|
'OpenStack Foundation', 'manual'),
|
||||||
]
|
]
|
||||||
|
|
||||||
# If false, no module index is generated.
|
# If false, no module index is generated.
|
||||||
|
|||||||
@@ -372,7 +372,7 @@ You can configure and install Ceilometer by following the documentation below :
|
|||||||
#. https://docs.openstack.org/ceilometer/latest
|
#. https://docs.openstack.org/ceilometer/latest
|
||||||
|
|
||||||
The built-in strategy 'basic_consolidation' provided by watcher requires
|
The built-in strategy 'basic_consolidation' provided by watcher requires
|
||||||
"**compute.node.cpu.percent**" and "**cpu_util**" measurements to be collected
|
"**compute.node.cpu.percent**" and "**cpu**" measurements to be collected
|
||||||
by Ceilometer.
|
by Ceilometer.
|
||||||
The measurements available depend on the hypervisors that OpenStack manages on
|
The measurements available depend on the hypervisors that OpenStack manages on
|
||||||
the specific implementation.
|
the specific implementation.
|
||||||
|
|||||||
@@ -47,6 +47,8 @@ unavailable as well as `instance_l3_cpu_cache`::
|
|||||||
[[local|localrc]]
|
[[local|localrc]]
|
||||||
enable_plugin watcher https://opendev.org/openstack/watcher
|
enable_plugin watcher https://opendev.org/openstack/watcher
|
||||||
|
|
||||||
|
enable_plugin watcher-dashboard https://opendev.org/openstack/watcher-dashboard
|
||||||
|
|
||||||
enable_plugin ceilometer https://opendev.org/openstack/ceilometer.git
|
enable_plugin ceilometer https://opendev.org/openstack/ceilometer.git
|
||||||
CEILOMETER_BACKEND=gnocchi
|
CEILOMETER_BACKEND=gnocchi
|
||||||
|
|
||||||
|
|||||||
@@ -56,9 +56,6 @@ Here is an example showing how you can write a plugin called ``NewStrategy``:
|
|||||||
# filepath: thirdparty/new.py
|
# filepath: thirdparty/new.py
|
||||||
# import path: thirdparty.new
|
# import path: thirdparty.new
|
||||||
import abc
|
import abc
|
||||||
|
|
||||||
import six
|
|
||||||
|
|
||||||
from watcher._i18n import _
|
from watcher._i18n import _
|
||||||
from watcher.decision_engine.strategy.strategies import base
|
from watcher.decision_engine.strategy.strategies import base
|
||||||
|
|
||||||
@@ -303,6 +300,6 @@ Using that you can now query the values for that specific metric:
|
|||||||
.. code-block:: py
|
.. code-block:: py
|
||||||
|
|
||||||
avg_meter = self.datasource_backend.statistic_aggregation(
|
avg_meter = self.datasource_backend.statistic_aggregation(
|
||||||
instance.uuid, 'cpu_util', self.periods['instance'],
|
instance.uuid, 'instance_cpu_usage', self.periods['instance'],
|
||||||
self.granularity,
|
self.granularity,
|
||||||
aggregation=self.aggregation_method['instance'])
|
aggregation=self.aggregation_method['instance'])
|
||||||
|
|||||||
@@ -26,8 +26,7 @@ metric service name plugins comment
|
|||||||
``compute_monitors`` option
|
``compute_monitors`` option
|
||||||
to ``cpu.virt_driver`` in
|
to ``cpu.virt_driver`` in
|
||||||
the nova.conf.
|
the nova.conf.
|
||||||
``cpu_util`` ceilometer_ none cpu_util has been removed
|
``cpu`` ceilometer_ none
|
||||||
since Stein.
|
|
||||||
============================ ============ ======= ===========================
|
============================ ============ ======= ===========================
|
||||||
|
|
||||||
.. _ceilometer: https://docs.openstack.org/ceilometer/latest/admin/telemetry-measurements.html#openstack-compute
|
.. _ceilometer: https://docs.openstack.org/ceilometer/latest/admin/telemetry-measurements.html#openstack-compute
|
||||||
|
|||||||
@@ -89,9 +89,9 @@ step 2: Create audit to do optimization
|
|||||||
.. code-block:: shell
|
.. code-block:: shell
|
||||||
|
|
||||||
$ openstack optimize audittemplate create \
|
$ openstack optimize audittemplate create \
|
||||||
at1 saving_energy --strategy saving_energy
|
saving_energy_template1 saving_energy --strategy saving_energy
|
||||||
|
|
||||||
$ openstack optimize audit create -a at1 \
|
$ openstack optimize audit create -a saving_energy_audit1 \
|
||||||
-p free_used_percent=20.0
|
-p free_used_percent=20.0
|
||||||
|
|
||||||
External Links
|
External Links
|
||||||
|
|||||||
@@ -22,14 +22,19 @@ The *vm_workload_consolidation* strategy requires the following metrics:
|
|||||||
============================ ============ ======= =========================
|
============================ ============ ======= =========================
|
||||||
metric service name plugins comment
|
metric service name plugins comment
|
||||||
============================ ============ ======= =========================
|
============================ ============ ======= =========================
|
||||||
``cpu_util`` ceilometer_ none cpu_util has been removed
|
``cpu`` ceilometer_ none
|
||||||
since Stein.
|
|
||||||
``memory.resident`` ceilometer_ none
|
``memory.resident`` ceilometer_ none
|
||||||
``memory`` ceilometer_ none
|
``memory`` ceilometer_ none
|
||||||
``disk.root.size`` ceilometer_ none
|
``disk.root.size`` ceilometer_ none
|
||||||
|
``compute.node.cpu.percent`` ceilometer_ none (optional) need to set the
|
||||||
|
``compute_monitors`` option
|
||||||
|
to ``cpu.virt_driver`` in the
|
||||||
|
nova.conf.
|
||||||
|
``hardware.memory.used`` ceilometer_ SNMP_ (optional)
|
||||||
============================ ============ ======= =========================
|
============================ ============ ======= =========================
|
||||||
|
|
||||||
.. _ceilometer: https://docs.openstack.org/ceilometer/latest/admin/telemetry-measurements.html#openstack-compute
|
.. _ceilometer: https://docs.openstack.org/ceilometer/latest/admin/telemetry-measurements.html#openstack-compute
|
||||||
|
.. _SNMP: https://docs.openstack.org/ceilometer/latest/admin/telemetry-measurements.html#snmp-based-meters
|
||||||
|
|
||||||
Cluster data model
|
Cluster data model
|
||||||
******************
|
******************
|
||||||
|
|||||||
@@ -27,9 +27,8 @@ metric service name plugins comment
|
|||||||
to ``cpu.virt_driver`` in the
|
to ``cpu.virt_driver`` in the
|
||||||
nova.conf.
|
nova.conf.
|
||||||
``hardware.memory.used`` ceilometer_ SNMP_
|
``hardware.memory.used`` ceilometer_ SNMP_
|
||||||
``cpu_util`` ceilometer_ none cpu_util has been removed
|
``cpu`` ceilometer_ none
|
||||||
since Stein.
|
``instance_ram_usage`` ceilometer_ none
|
||||||
``memory.resident`` ceilometer_ none
|
|
||||||
============================ ============ ======= =============================
|
============================ ============ ======= =============================
|
||||||
|
|
||||||
.. _ceilometer: https://docs.openstack.org/ceilometer/latest/admin/telemetry-measurements.html#openstack-compute
|
.. _ceilometer: https://docs.openstack.org/ceilometer/latest/admin/telemetry-measurements.html#openstack-compute
|
||||||
@@ -107,10 +106,10 @@ parameter type default Value description
|
|||||||
period of all received ones.
|
period of all received ones.
|
||||||
==================== ====== ===================== =============================
|
==================== ====== ===================== =============================
|
||||||
|
|
||||||
.. |metrics| replace:: ["cpu_util", "memory.resident"]
|
.. |metrics| replace:: ["instance_cpu_usage", "instance_ram_usage"]
|
||||||
.. |thresholds| replace:: {"cpu_util": 0.2, "memory.resident": 0.2}
|
.. |thresholds| replace:: {"instance_cpu_usage": 0.2, "instance_ram_usage": 0.2}
|
||||||
.. |weights| replace:: {"cpu_util_weight": 1.0, "memory.resident_weight": 1.0}
|
.. |weights| replace:: {"instance_cpu_usage_weight": 1.0, "instance_ram_usage_weight": 1.0}
|
||||||
.. |instance_metrics| replace:: {"cpu_util": "compute.node.cpu.percent", "memory.resident": "hardware.memory.used"}
|
.. |instance_metrics| replace:: {"instance_cpu_usage": "compute.node.cpu.percent", "instance_ram_usage": "hardware.memory.used"}
|
||||||
.. |periods| replace:: {"instance": 720, "node": 600}
|
.. |periods| replace:: {"instance": 720, "node": 600}
|
||||||
|
|
||||||
Efficacy Indicator
|
Efficacy Indicator
|
||||||
@@ -136,8 +135,8 @@ How to use it ?
|
|||||||
at1 workload_balancing --strategy workload_stabilization
|
at1 workload_balancing --strategy workload_stabilization
|
||||||
|
|
||||||
$ openstack optimize audit create -a at1 \
|
$ openstack optimize audit create -a at1 \
|
||||||
-p thresholds='{"memory.resident": 0.05}' \
|
-p thresholds='{"instance_ram_usage": 0.05}' \
|
||||||
-p metrics='["memory.resident"]'
|
-p metrics='["instance_ram_usage"]'
|
||||||
|
|
||||||
External Links
|
External Links
|
||||||
--------------
|
--------------
|
||||||
|
|||||||
@@ -24,8 +24,7 @@ The *workload_balance* strategy requires the following metrics:
|
|||||||
======================= ============ ======= =========================
|
======================= ============ ======= =========================
|
||||||
metric service name plugins comment
|
metric service name plugins comment
|
||||||
======================= ============ ======= =========================
|
======================= ============ ======= =========================
|
||||||
``cpu_util`` ceilometer_ none cpu_util has been removed
|
``cpu`` ceilometer_ none
|
||||||
since Stein.
|
|
||||||
``memory.resident`` ceilometer_ none
|
``memory.resident`` ceilometer_ none
|
||||||
======================= ============ ======= =========================
|
======================= ============ ======= =========================
|
||||||
|
|
||||||
@@ -65,15 +64,16 @@ Configuration
|
|||||||
|
|
||||||
Strategy parameters are:
|
Strategy parameters are:
|
||||||
|
|
||||||
============== ====== ============= ====================================
|
============== ====== ==================== ====================================
|
||||||
parameter type default Value description
|
parameter type default Value description
|
||||||
============== ====== ============= ====================================
|
============== ====== ==================== ====================================
|
||||||
``metrics`` String 'cpu_util' Workload balance base on cpu or ram
|
``metrics`` String 'instance_cpu_usage' Workload balance base on cpu or ram
|
||||||
utilization. choice: ['cpu_util',
|
utilization. Choices:
|
||||||
'memory.resident']
|
['instance_cpu_usage',
|
||||||
``threshold`` Number 25.0 Workload threshold for migration
|
'instance_ram_usage']
|
||||||
``period`` Number 300 Aggregate time period of ceilometer
|
``threshold`` Number 25.0 Workload threshold for migration
|
||||||
============== ====== ============= ====================================
|
``period`` Number 300 Aggregate time period of ceilometer
|
||||||
|
============== ====== ==================== ====================================
|
||||||
|
|
||||||
Efficacy Indicator
|
Efficacy Indicator
|
||||||
------------------
|
------------------
|
||||||
@@ -95,7 +95,7 @@ How to use it ?
|
|||||||
at1 workload_balancing --strategy workload_balance
|
at1 workload_balancing --strategy workload_balance
|
||||||
|
|
||||||
$ openstack optimize audit create -a at1 -p threshold=26.0 \
|
$ openstack optimize audit create -a at1 -p threshold=26.0 \
|
||||||
-p period=310 -p metrics=cpu_util
|
-p period=310 -p metrics=instance_cpu_usage
|
||||||
|
|
||||||
External Links
|
External Links
|
||||||
--------------
|
--------------
|
||||||
|
|||||||
@@ -1,161 +0,0 @@
|
|||||||
alabaster==0.7.10
|
|
||||||
alembic==0.9.8
|
|
||||||
amqp==2.2.2
|
|
||||||
appdirs==1.4.3
|
|
||||||
APScheduler==3.5.1
|
|
||||||
asn1crypto==0.24.0
|
|
||||||
automaton==1.14.0
|
|
||||||
Babel==2.5.3
|
|
||||||
beautifulsoup4==4.6.0
|
|
||||||
cachetools==2.0.1
|
|
||||||
certifi==2018.1.18
|
|
||||||
cffi==1.11.5
|
|
||||||
chardet==3.0.4
|
|
||||||
cliff==2.11.0
|
|
||||||
cmd2==0.8.1
|
|
||||||
contextlib2==0.5.5
|
|
||||||
coverage==4.5.1
|
|
||||||
croniter==0.3.20
|
|
||||||
cryptography==2.1.4
|
|
||||||
debtcollector==1.19.0
|
|
||||||
decorator==4.2.1
|
|
||||||
deprecation==2.0
|
|
||||||
doc8==0.8.0
|
|
||||||
docutils==0.14
|
|
||||||
dogpile.cache==0.6.5
|
|
||||||
dulwich==0.19.0
|
|
||||||
enum34==1.1.6
|
|
||||||
enum-compat==0.0.2
|
|
||||||
eventlet==0.20.0
|
|
||||||
extras==1.0.0
|
|
||||||
fasteners==0.14.1
|
|
||||||
fixtures==3.0.0
|
|
||||||
freezegun==0.3.10
|
|
||||||
future==0.16.0
|
|
||||||
futurist==1.8.0
|
|
||||||
gitdb2==2.0.3
|
|
||||||
GitPython==2.1.8
|
|
||||||
gnocchiclient==7.0.1
|
|
||||||
greenlet==0.4.13
|
|
||||||
idna==2.6
|
|
||||||
imagesize==1.0.0
|
|
||||||
iso8601==0.1.12
|
|
||||||
Jinja2==2.10
|
|
||||||
jmespath==0.9.3
|
|
||||||
jsonpatch==1.21
|
|
||||||
jsonpointer==2.0
|
|
||||||
jsonschema==2.6.0
|
|
||||||
keystoneauth1==3.4.0
|
|
||||||
keystonemiddleware==4.21.0
|
|
||||||
kombu==4.1.0
|
|
||||||
linecache2==1.0.0
|
|
||||||
logutils==0.3.5
|
|
||||||
lxml==4.1.1
|
|
||||||
Mako==1.0.7
|
|
||||||
MarkupSafe==1.0
|
|
||||||
mccabe==0.2.1
|
|
||||||
microversion_parse==0.2.1
|
|
||||||
mock==2.0.0
|
|
||||||
monotonic==1.4
|
|
||||||
mox3==0.25.0
|
|
||||||
msgpack==0.5.6
|
|
||||||
munch==2.2.0
|
|
||||||
netaddr==0.7.19
|
|
||||||
netifaces==0.10.6
|
|
||||||
networkx==2.2
|
|
||||||
openstackdocstheme==1.20.0
|
|
||||||
openstacksdk==0.12.0
|
|
||||||
os-api-ref===1.4.0
|
|
||||||
os-client-config==1.29.0
|
|
||||||
os-service-types==1.2.0
|
|
||||||
os-testr==1.0.0
|
|
||||||
osc-lib==1.10.0
|
|
||||||
os-resource-classes==0.4.0
|
|
||||||
oslo.cache==1.29.0
|
|
||||||
oslo.concurrency==3.26.0
|
|
||||||
oslo.config==5.2.0
|
|
||||||
oslo.context==2.21.0
|
|
||||||
oslo.db==4.35.0
|
|
||||||
oslo.i18n==3.20.0
|
|
||||||
oslo.log==3.37.0
|
|
||||||
oslo.messaging==8.1.2
|
|
||||||
oslo.middleware==3.35.0
|
|
||||||
oslo.policy==1.34.0
|
|
||||||
oslo.reports==1.27.0
|
|
||||||
oslo.serialization==2.25.0
|
|
||||||
oslo.service==1.30.0
|
|
||||||
oslo.upgradecheck==0.1.0
|
|
||||||
oslo.utils==3.36.0
|
|
||||||
oslo.versionedobjects==1.32.0
|
|
||||||
oslotest==3.3.0
|
|
||||||
packaging==17.1
|
|
||||||
Paste==2.0.3
|
|
||||||
PasteDeploy==1.5.2
|
|
||||||
pbr==3.1.1
|
|
||||||
pecan==1.3.2
|
|
||||||
pika==0.10.0
|
|
||||||
pika-pool==0.1.3
|
|
||||||
prettytable==0.7.2
|
|
||||||
psutil==5.4.3
|
|
||||||
pycadf==2.7.0
|
|
||||||
pycparser==2.18
|
|
||||||
Pygments==2.2.0
|
|
||||||
pyinotify==0.9.6
|
|
||||||
pyOpenSSL==17.5.0
|
|
||||||
pyparsing==2.2.0
|
|
||||||
pyperclip==1.6.0
|
|
||||||
python-ceilometerclient==2.9.0
|
|
||||||
python-cinderclient==3.5.0
|
|
||||||
python-dateutil==2.7.0
|
|
||||||
python-editor==1.0.3
|
|
||||||
python-glanceclient==2.9.1
|
|
||||||
python-ironicclient==2.5.0
|
|
||||||
python-keystoneclient==3.15.0
|
|
||||||
python-mimeparse==1.6.0
|
|
||||||
python-monascaclient==1.12.0
|
|
||||||
python-neutronclient==6.7.0
|
|
||||||
python-novaclient==14.1.0
|
|
||||||
python-openstackclient==3.14.0
|
|
||||||
python-subunit==1.2.0
|
|
||||||
pytz==2018.3
|
|
||||||
PyYAML==3.12
|
|
||||||
reno==2.7.0
|
|
||||||
repoze.lru==0.7
|
|
||||||
requests==2.18.4
|
|
||||||
requestsexceptions==1.4.0
|
|
||||||
restructuredtext-lint==1.1.3
|
|
||||||
rfc3986==1.1.0
|
|
||||||
Routes==2.4.1
|
|
||||||
simplegeneric==0.8.1
|
|
||||||
simplejson==3.13.2
|
|
||||||
six==1.11.0
|
|
||||||
smmap2==2.0.3
|
|
||||||
snowballstemmer==1.2.1
|
|
||||||
Sphinx==1.6.5
|
|
||||||
sphinxcontrib-httpdomain==1.6.1
|
|
||||||
sphinxcontrib-pecanwsme==0.8.0
|
|
||||||
sphinxcontrib-websupport==1.0.1
|
|
||||||
SQLAlchemy==1.2.5
|
|
||||||
sqlalchemy-migrate==0.11.0
|
|
||||||
sqlparse==0.2.4
|
|
||||||
statsd==3.2.2
|
|
||||||
stestr==2.0.0
|
|
||||||
stevedore==1.28.0
|
|
||||||
taskflow==3.7.1
|
|
||||||
Tempita==0.5.2
|
|
||||||
tenacity==4.9.0
|
|
||||||
testresources==2.0.1
|
|
||||||
testscenarios==0.5.0
|
|
||||||
testtools==2.3.0
|
|
||||||
traceback2==1.4.0
|
|
||||||
tzlocal==1.5.1
|
|
||||||
ujson==1.35
|
|
||||||
unittest2==1.1.0
|
|
||||||
urllib3==1.22
|
|
||||||
vine==1.1.4
|
|
||||||
waitress==1.1.0
|
|
||||||
warlock==1.3.0
|
|
||||||
WebOb==1.8.5
|
|
||||||
WebTest==2.0.29
|
|
||||||
wrapt==1.10.11
|
|
||||||
WSME==0.9.2
|
|
||||||
@@ -1,15 +0,0 @@
|
|||||||
- hosts: primary
|
|
||||||
tasks:
|
|
||||||
|
|
||||||
- name: Copy files from {{ ansible_user_dir }}/workspace/ on node
|
|
||||||
synchronize:
|
|
||||||
src: '{{ ansible_user_dir }}/workspace/'
|
|
||||||
dest: '{{ zuul.executor.log_root }}'
|
|
||||||
mode: pull
|
|
||||||
copy_links: true
|
|
||||||
verify_host: true
|
|
||||||
rsync_opts:
|
|
||||||
- --include=/logs/**
|
|
||||||
- --include=*/
|
|
||||||
- --exclude=*
|
|
||||||
- --prune-empty-dirs
|
|
||||||
@@ -1,60 +0,0 @@
|
|||||||
- hosts: all
|
|
||||||
name: legacy-grenade-dsvm-watcher
|
|
||||||
tasks:
|
|
||||||
|
|
||||||
- name: Ensure legacy workspace directory
|
|
||||||
file:
|
|
||||||
path: '{{ ansible_user_dir }}/workspace'
|
|
||||||
state: directory
|
|
||||||
|
|
||||||
- shell:
|
|
||||||
cmd: |
|
|
||||||
set -e
|
|
||||||
set -x
|
|
||||||
cat > clonemap.yaml << EOF
|
|
||||||
clonemap:
|
|
||||||
- name: openstack/devstack-gate
|
|
||||||
dest: devstack-gate
|
|
||||||
EOF
|
|
||||||
/usr/zuul-env/bin/zuul-cloner -m clonemap.yaml --cache-dir /opt/git \
|
|
||||||
https://opendev.org \
|
|
||||||
openstack/devstack-gate
|
|
||||||
executable: /bin/bash
|
|
||||||
chdir: '{{ ansible_user_dir }}/workspace'
|
|
||||||
environment: '{{ zuul | zuul_legacy_vars }}'
|
|
||||||
|
|
||||||
- shell:
|
|
||||||
cmd: |
|
|
||||||
set -e
|
|
||||||
set -x
|
|
||||||
export PYTHONUNBUFFERED=true
|
|
||||||
|
|
||||||
export PROJECTS="openstack/grenade $PROJECTS"
|
|
||||||
export PROJECTS="openstack/watcher $PROJECTS"
|
|
||||||
export PROJECTS="openstack/watcher-tempest-plugin $PROJECTS"
|
|
||||||
export PROJECTS="openstack/python-watcherclient $PROJECTS"
|
|
||||||
export DEVSTACK_PROJECT_FROM_GIT="python-watcherclient $DEVSTACK_PROJECT_FROM_GIT"
|
|
||||||
|
|
||||||
export GRENADE_PLUGINRC="enable_grenade_plugin watcher https://opendev.org/openstack/watcher"
|
|
||||||
export DEVSTACK_LOCAL_CONFIG+=$'\n'"export TEMPEST_PLUGINS='/opt/stack/new/watcher-tempest-plugin'"
|
|
||||||
|
|
||||||
export DEVSTACK_GATE_TEMPEST_NOTESTS=1
|
|
||||||
export DEVSTACK_GATE_GRENADE=pullup
|
|
||||||
export DEVSTACK_GATE_USE_PYTHON3=True
|
|
||||||
export BRANCH_OVERRIDE=default
|
|
||||||
if [ "$BRANCH_OVERRIDE" != "default" ] ; then
|
|
||||||
export OVERRIDE_ZUUL_BRANCH=$BRANCH_OVERRIDE
|
|
||||||
fi
|
|
||||||
# Add configuration values for enabling security features in local.conf
|
|
||||||
function pre_test_hook {
|
|
||||||
if [ -f /opt/stack/old/watcher-tempest-plugin/tools/pre_test_hook.sh ] ; then
|
|
||||||
. /opt/stack/old/watcher-tempest-plugin/tools/pre_test_hook.sh
|
|
||||||
fi
|
|
||||||
}
|
|
||||||
export -f pre_test_hook
|
|
||||||
|
|
||||||
cp devstack-gate/devstack-vm-gate-wrap.sh ./safe-devstack-vm-gate-wrap.sh
|
|
||||||
./safe-devstack-vm-gate-wrap.sh
|
|
||||||
executable: /bin/bash
|
|
||||||
chdir: '{{ ansible_user_dir }}/workspace'
|
|
||||||
environment: '{{ zuul | zuul_legacy_vars }}'
|
|
||||||
@@ -0,0 +1,20 @@
|
|||||||
|
---
|
||||||
|
upgrade:
|
||||||
|
- |
|
||||||
|
The default value of ``[oslo_policy] policy_file`` config option has
|
||||||
|
been changed from ``policy.json`` to ``policy.yaml``.
|
||||||
|
Operators who are utilizing customized or previously generated
|
||||||
|
static policy JSON files (which are not needed by default), should
|
||||||
|
generate new policy files or convert them in YAML format. Use the
|
||||||
|
`oslopolicy-convert-json-to-yaml
|
||||||
|
<https://docs.openstack.org/oslo.policy/latest/cli/oslopolicy-convert-json-to-yaml.html>`_
|
||||||
|
tool to convert a JSON to YAML formatted policy file in
|
||||||
|
backward compatible way.
|
||||||
|
deprecations:
|
||||||
|
- |
|
||||||
|
Use of JSON policy files was deprecated by the ``oslo.policy`` library
|
||||||
|
during the Victoria development cycle. As a result, this deprecation is
|
||||||
|
being noted in the Wallaby cycle with an anticipated future removal of support
|
||||||
|
by ``oslo.policy``. As such operators will need to convert to YAML policy
|
||||||
|
files. Please see the upgrade notes for details on migration of any
|
||||||
|
custom policy files.
|
||||||
6
releasenotes/source/2023.1.rst
Normal file
6
releasenotes/source/2023.1.rst
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
===========================
|
||||||
|
2023.1 Series Release Notes
|
||||||
|
===========================
|
||||||
|
|
||||||
|
.. release-notes::
|
||||||
|
:branch: stable/2023.1
|
||||||
6
releasenotes/source/2023.2.rst
Normal file
6
releasenotes/source/2023.2.rst
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
===========================
|
||||||
|
2023.2 Series Release Notes
|
||||||
|
===========================
|
||||||
|
|
||||||
|
.. release-notes::
|
||||||
|
:branch: stable/2023.2
|
||||||
@@ -53,8 +53,7 @@ source_suffix = '.rst'
|
|||||||
master_doc = 'index'
|
master_doc = 'index'
|
||||||
|
|
||||||
# General information about the project.
|
# General information about the project.
|
||||||
project = u'watcher'
|
copyright = '2016, Watcher developers'
|
||||||
copyright = u'2016, Watcher developers'
|
|
||||||
|
|
||||||
# Release notes are version independent
|
# Release notes are version independent
|
||||||
# The short X.Y version.
|
# The short X.Y version.
|
||||||
@@ -91,11 +90,15 @@ exclude_patterns = ['_build']
|
|||||||
#show_authors = False
|
#show_authors = False
|
||||||
|
|
||||||
# The name of the Pygments (syntax highlighting) style to use.
|
# The name of the Pygments (syntax highlighting) style to use.
|
||||||
pygments_style = 'sphinx'
|
pygments_style = 'native'
|
||||||
|
|
||||||
# A list of ignored prefixes for module index sorting.
|
# A list of ignored prefixes for module index sorting.
|
||||||
#modindex_common_prefix = []
|
#modindex_common_prefix = []
|
||||||
|
|
||||||
|
# openstackdocstheme options
|
||||||
|
openstackdocs_repo_name = 'openstack/watcher'
|
||||||
|
openstackdocs_bug_project = 'watcher'
|
||||||
|
openstackdocs_bug_tag = ''
|
||||||
|
|
||||||
# -- Options for HTML output --------------------------------------------------
|
# -- Options for HTML output --------------------------------------------------
|
||||||
|
|
||||||
@@ -193,8 +196,8 @@ latex_elements = {
|
|||||||
# Grouping the document tree into LaTeX files. List of tuples
|
# Grouping the document tree into LaTeX files. List of tuples
|
||||||
# (source start file, target name, title, author, documentclass [howto/manual])
|
# (source start file, target name, title, author, documentclass [howto/manual])
|
||||||
latex_documents = [
|
latex_documents = [
|
||||||
('index', 'watcher.tex', u'Watcher Documentation',
|
('index', 'watcher.tex', 'Watcher Documentation',
|
||||||
u'Watcher developers', 'manual'),
|
'Watcher developers', 'manual'),
|
||||||
]
|
]
|
||||||
|
|
||||||
# The name of an image file (relative to this directory) to place at the top of
|
# The name of an image file (relative to this directory) to place at the top of
|
||||||
@@ -223,8 +226,8 @@ latex_documents = [
|
|||||||
# One entry per manual page. List of tuples
|
# One entry per manual page. List of tuples
|
||||||
# (source start file, name, description, authors, manual section).
|
# (source start file, name, description, authors, manual section).
|
||||||
man_pages = [
|
man_pages = [
|
||||||
('index', 'watcher', u'Watcher Documentation',
|
('index', 'watcher', 'Watcher Documentation',
|
||||||
[u'Watcher developers'], 1)
|
['Watcher developers'], 1)
|
||||||
]
|
]
|
||||||
|
|
||||||
# If true, show URL addresses after external links.
|
# If true, show URL addresses after external links.
|
||||||
@@ -237,8 +240,8 @@ man_pages = [
|
|||||||
# (source start file, target name, title, author,
|
# (source start file, target name, title, author,
|
||||||
# dir menu entry, description, category)
|
# dir menu entry, description, category)
|
||||||
texinfo_documents = [
|
texinfo_documents = [
|
||||||
('index', 'watcher', u'Watcher Documentation',
|
('index', 'watcher', 'Watcher Documentation',
|
||||||
u'Watcher developers', 'watcher', 'One line description of project.',
|
'Watcher developers', 'watcher', 'One line description of project.',
|
||||||
'Miscellaneous'),
|
'Miscellaneous'),
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|||||||
@@ -21,6 +21,14 @@ Contents:
|
|||||||
:maxdepth: 1
|
:maxdepth: 1
|
||||||
|
|
||||||
unreleased
|
unreleased
|
||||||
|
2023.2
|
||||||
|
2023.1
|
||||||
|
zed
|
||||||
|
yoga
|
||||||
|
xena
|
||||||
|
wallaby
|
||||||
|
victoria
|
||||||
|
ussuri
|
||||||
train
|
train
|
||||||
stein
|
stein
|
||||||
rocky
|
rocky
|
||||||
|
|||||||
@@ -1,14 +1,17 @@
|
|||||||
# Andi Chandler <andi@gowling.com>, 2017. #zanata
|
# Andi Chandler <andi@gowling.com>, 2017. #zanata
|
||||||
# Andi Chandler <andi@gowling.com>, 2018. #zanata
|
# Andi Chandler <andi@gowling.com>, 2018. #zanata
|
||||||
|
# Andi Chandler <andi@gowling.com>, 2020. #zanata
|
||||||
|
# Andi Chandler <andi@gowling.com>, 2022. #zanata
|
||||||
|
# Andi Chandler <andi@gowling.com>, 2023. #zanata
|
||||||
msgid ""
|
msgid ""
|
||||||
msgstr ""
|
msgstr ""
|
||||||
"Project-Id-Version: python-watcher\n"
|
"Project-Id-Version: python-watcher\n"
|
||||||
"Report-Msgid-Bugs-To: \n"
|
"Report-Msgid-Bugs-To: \n"
|
||||||
"POT-Creation-Date: 2018-11-08 01:22+0000\n"
|
"POT-Creation-Date: 2023-08-14 03:05+0000\n"
|
||||||
"MIME-Version: 1.0\n"
|
"MIME-Version: 1.0\n"
|
||||||
"Content-Type: text/plain; charset=UTF-8\n"
|
"Content-Type: text/plain; charset=UTF-8\n"
|
||||||
"Content-Transfer-Encoding: 8bit\n"
|
"Content-Transfer-Encoding: 8bit\n"
|
||||||
"PO-Revision-Date: 2018-11-07 06:15+0000\n"
|
"PO-Revision-Date: 2023-06-21 07:54+0000\n"
|
||||||
"Last-Translator: Andi Chandler <andi@gowling.com>\n"
|
"Last-Translator: Andi Chandler <andi@gowling.com>\n"
|
||||||
"Language-Team: English (United Kingdom)\n"
|
"Language-Team: English (United Kingdom)\n"
|
||||||
"Language: en_GB\n"
|
"Language: en_GB\n"
|
||||||
@@ -54,6 +57,67 @@ msgstr "1.7.0"
|
|||||||
msgid "1.9.0"
|
msgid "1.9.0"
|
||||||
msgstr "1.9.0"
|
msgstr "1.9.0"
|
||||||
|
|
||||||
|
msgid "2.0.0"
|
||||||
|
msgstr "2.0.0"
|
||||||
|
|
||||||
|
msgid "2023.1 Series Release Notes"
|
||||||
|
msgstr "2023.1 Series Release Notes"
|
||||||
|
|
||||||
|
msgid "3.0.0"
|
||||||
|
msgstr "3.0.0"
|
||||||
|
|
||||||
|
msgid "4.0.0"
|
||||||
|
msgstr "4.0.0"
|
||||||
|
|
||||||
|
msgid "6.0.0"
|
||||||
|
msgstr "6.0.0"
|
||||||
|
|
||||||
|
msgid "A ``watcher-status upgrade check`` has been added for this."
|
||||||
|
msgstr "A ``watcher-status upgrade check`` has been added for this."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"A new threadpool for the decision engine that contributors can use to "
|
||||||
|
"improve the performance of many operations, primarily I/O bound onces. The "
|
||||||
|
"amount of workers used by the decision engine threadpool can be configured "
|
||||||
|
"to scale according to the available infrastructure using the "
|
||||||
|
"`watcher_decision_engine.max_general_workers` config option. Documentation "
|
||||||
|
"for contributors to effectively use this threadpool is available online: "
|
||||||
|
"https://docs.openstack.org/watcher/latest/contributor/concurrency.html"
|
||||||
|
msgstr ""
|
||||||
|
"A new threadpool for the decision engine that contributors can use to "
|
||||||
|
"improve the performance of many operations, primarily I/O bound onces. The "
|
||||||
|
"amount of workers used by the decision engine threadpool can be configured "
|
||||||
|
"to scale according to the available infrastructure using the "
|
||||||
|
"`watcher_decision_engine.max_general_workers` config option. Documentation "
|
||||||
|
"for contributors to effectively use this threadpool is available online: "
|
||||||
|
"https://docs.openstack.org/watcher/latest/contributor/concurrency.html"
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"API calls while building the Compute data model will be retried upon "
|
||||||
|
"failure. The amount of failures allowed before giving up and the time before "
|
||||||
|
"reattempting are configurable. The `api_call_retries` and "
|
||||||
|
"`api_query_timeout` parameters in the `[collector]` group can be used to "
|
||||||
|
"adjust these paremeters. 10 retries with a 1 second time in between "
|
||||||
|
"reattempts is the default."
|
||||||
|
msgstr ""
|
||||||
|
"API calls while building the Compute data model will be retried upon "
|
||||||
|
"failure. The amount of failures allowed before giving up and the time before "
|
||||||
|
"reattempting are configurable. The `api_call_retries` and "
|
||||||
|
"`api_query_timeout` parameters in the `[collector]` group can be used to "
|
||||||
|
"adjust these parameters. 10 retries with a 1 second time in between "
|
||||||
|
"reattempts is the default."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Add a new webhook API and a new audit type EVENT, the microversion is 1.4. "
|
||||||
|
"Now Watcher user can create audit with EVENT type and the audit will be "
|
||||||
|
"triggered by webhook API. The user guide is available online: https://docs."
|
||||||
|
"openstack.org/watcher/latest/user/event_type_audit.html"
|
||||||
|
msgstr ""
|
||||||
|
"Add a new webhook API and a new audit type EVENT, the microversion is 1.4. "
|
||||||
|
"Now Watcher user can create audit with EVENT type and the audit will be "
|
||||||
|
"triggered by webhook API. The user guide is available online: https://docs."
|
||||||
|
"openstack.org/watcher/latest/user/event_type_audit.html"
|
||||||
|
|
||||||
msgid "Add a service supervisor to watch Watcher deamons."
|
msgid "Add a service supervisor to watch Watcher deamons."
|
||||||
msgstr "Add a service supervisor to watch Watcher daemons."
|
msgstr "Add a service supervisor to watch Watcher daemons."
|
||||||
|
|
||||||
@@ -67,6 +131,24 @@ msgstr ""
|
|||||||
"Add description property for dynamic action. Admin can see detail "
|
"Add description property for dynamic action. Admin can see detail "
|
||||||
"information of any specify action."
|
"information of any specify action."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Add force field to Audit. User can set --force to enable the new option when "
|
||||||
|
"launching audit. If force is True, audit will be executed despite of ongoing "
|
||||||
|
"actionplan. The new audit may create a wrong actionplan if they use the same "
|
||||||
|
"data model."
|
||||||
|
msgstr ""
|
||||||
|
"Add force field to Audit. User can set --force to enable the new option when "
|
||||||
|
"launching audit. If force is True, audit will be executed despite of ongoing "
|
||||||
|
"actionplan. The new audit may create a wrong actionplan if they use the same "
|
||||||
|
"data model."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Add keystone_client Group for user to configure 'interface' and "
|
||||||
|
"'region_name' by watcher.conf. The default value of 'interface' is 'admin'."
|
||||||
|
msgstr ""
|
||||||
|
"Add keystone_client Group for user to configure 'interface' and "
|
||||||
|
"'region_name' by watcher.conf. The default value of 'interface' is 'admin'."
|
||||||
|
|
||||||
msgid "Add notifications related to Action object."
|
msgid "Add notifications related to Action object."
|
||||||
msgstr "Add notifications related to Action object."
|
msgstr "Add notifications related to Action object."
|
||||||
|
|
||||||
@@ -79,6 +161,25 @@ msgstr "Add notifications related to Audit object."
|
|||||||
msgid "Add notifications related to Service object."
|
msgid "Add notifications related to Service object."
|
||||||
msgstr "Add notifications related to Service object."
|
msgstr "Add notifications related to Service object."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Add show data model api for Watcher. New in version 1.3. User can use "
|
||||||
|
"'openstack optimize datamodel list' command to view the current data model "
|
||||||
|
"information in memory. User can also add '--audit <Audit_UUID>' to view "
|
||||||
|
"specific data model in memory filted by the scope in audit. User can also "
|
||||||
|
"add '--detail' to view detailed information about current data model. User "
|
||||||
|
"can also add '--type <type>' to specify the type of data model. Default type "
|
||||||
|
"is 'compute'. In the future, type 'storage' and 'baremetal' will be "
|
||||||
|
"supported."
|
||||||
|
msgstr ""
|
||||||
|
"Add show data model API for Watcher. New in version 1.3. User can use "
|
||||||
|
"'openstack optimize datamodel list' command to view the current data model "
|
||||||
|
"information in memory. User can also add '--audit <Audit_UUID>' to view "
|
||||||
|
"specific data model in memory filtered by the scope in audit. User can also "
|
||||||
|
"add '--detail' to view detailed information about current data model. User "
|
||||||
|
"can also add '--type <type>' to specify the type of data model. Default type "
|
||||||
|
"is 'compute'. In the future, type 'storage' and 'baremetal' will be "
|
||||||
|
"supported."
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"Add start_time and end_time fields in audits table. User can set the start "
|
"Add start_time and end_time fields in audits table. User can set the start "
|
||||||
"time and/or end time when creating CONTINUOUS audit."
|
"time and/or end time when creating CONTINUOUS audit."
|
||||||
@@ -93,6 +194,19 @@ msgstr ""
|
|||||||
"Add superseded state for an action plan if the cluster data model has "
|
"Add superseded state for an action plan if the cluster data model has "
|
||||||
"changed after it has been created."
|
"changed after it has been created."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Added Placement API helper to Watcher. Now Watcher can get information about "
|
||||||
|
"resource providers, it can be used for the data model and strategies. Config "
|
||||||
|
"group placement_client with options 'api_version', 'interface' and "
|
||||||
|
"'region_name' is also added. The default values for 'api_version' and "
|
||||||
|
"'interface' are 1.29 and 'public', respectively."
|
||||||
|
msgstr ""
|
||||||
|
"Added Placement API helper to Watcher. Now Watcher can get information about "
|
||||||
|
"resource providers, it can be used for the data model and strategies. Config "
|
||||||
|
"group placement_client with options 'api_version', 'interface' and "
|
||||||
|
"'region_name' is also added. The default values for 'api_version' and "
|
||||||
|
"'interface' are 1.29 and 'public', respectively."
|
||||||
|
|
||||||
msgid "Added SUSPENDED audit state"
|
msgid "Added SUSPENDED audit state"
|
||||||
msgstr "Added SUSPENDED audit state"
|
msgstr "Added SUSPENDED audit state"
|
||||||
|
|
||||||
@@ -107,6 +221,31 @@ msgstr ""
|
|||||||
"scoring engine by different Strategies, which improve the code and data "
|
"scoring engine by different Strategies, which improve the code and data "
|
||||||
"model re-use."
|
"model re-use."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Added a new config option 'action_execution_rule' which is a dict type. Its "
|
||||||
|
"key field is strategy name and the value is 'ALWAYS' or 'ANY'. 'ALWAYS' "
|
||||||
|
"means the callback function returns True as usual. 'ANY' means the return "
|
||||||
|
"depends on the result of previous action execution. The callback returns "
|
||||||
|
"True if previous action gets failed, and the engine continues to run the "
|
||||||
|
"next action. If previous action executes success, the callback returns False "
|
||||||
|
"then the next action will be ignored. For strategies that aren't in "
|
||||||
|
"'action_execution_rule', the callback always returns True. Please add the "
|
||||||
|
"next section in the watcher.conf file if your strategy needs this feature. "
|
||||||
|
"[watcher_workflow_engines.taskflow] action_execution_rule = {'your strategy "
|
||||||
|
"name': 'ANY'}"
|
||||||
|
msgstr ""
|
||||||
|
"Added a new config option 'action_execution_rule' which is a dict type. Its "
|
||||||
|
"key field is strategy name and the value is 'ALWAYS' or 'ANY'. 'ALWAYS' "
|
||||||
|
"means the callback function returns True as usual. 'ANY' means the return "
|
||||||
|
"depends on the result of previous action execution. The callback returns "
|
||||||
|
"True if previous action gets failed, and the engine continues to run the "
|
||||||
|
"next action. If previous action executes success, the callback returns False "
|
||||||
|
"then the next action will be ignored. For strategies that aren't in "
|
||||||
|
"'action_execution_rule', the callback always returns True. Please add the "
|
||||||
|
"next section in the watcher.conf file if your strategy needs this feature. "
|
||||||
|
"[watcher_workflow_engines.taskflow] action_execution_rule = {'your strategy "
|
||||||
|
"name': 'ANY'}"
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"Added a new strategy based on the airflow of servers. This strategy makes "
|
"Added a new strategy based on the airflow of servers. This strategy makes "
|
||||||
"decisions to migrate VMs to make the airflow uniform."
|
"decisions to migrate VMs to make the airflow uniform."
|
||||||
@@ -248,6 +387,15 @@ msgstr ""
|
|||||||
"The strategy migrates many instances and volumes efficiently with minimum "
|
"The strategy migrates many instances and volumes efficiently with minimum "
|
||||||
"downtime automatically."
|
"downtime automatically."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Added strategy \"node resource consolidation\". This strategy is used to "
|
||||||
|
"centralize VMs to as few nodes as possible by VM migration. User can set an "
|
||||||
|
"input parameter to decide how to select the destination node."
|
||||||
|
msgstr ""
|
||||||
|
"Added strategy \"node resource consolidation\". This strategy is used to "
|
||||||
|
"centralize VMs to as few nodes as possible by VM migration. User can set an "
|
||||||
|
"input parameter to decide how to select the destination node."
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"Added strategy to identify and migrate a Noisy Neighbor - a low priority VM "
|
"Added strategy to identify and migrate a Noisy Neighbor - a low priority VM "
|
||||||
"that negatively affects peformance of a high priority VM by over utilizing "
|
"that negatively affects peformance of a high priority VM by over utilizing "
|
||||||
@@ -284,6 +432,19 @@ msgstr ""
|
|||||||
msgid "Adds baremetal data model in Watcher"
|
msgid "Adds baremetal data model in Watcher"
|
||||||
msgstr "Adds baremetal data model in Watcher"
|
msgstr "Adds baremetal data model in Watcher"
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"All datasources can now be configured to retry retrieving a metric upon "
|
||||||
|
"encountering an error. Between each attempt will be a set amount of time "
|
||||||
|
"which can be adjusted from the configuration. These configuration options "
|
||||||
|
"can be found in the `[watcher_datasources]` group and are named "
|
||||||
|
"`query_max_retries` and `query_timeout`."
|
||||||
|
msgstr ""
|
||||||
|
"All datasources can now be configured to retry retrieving a metric upon "
|
||||||
|
"encountering an error. Between each attempt will be a set amount of time "
|
||||||
|
"which can be adjusted from the configuration. These configuration options "
|
||||||
|
"can be found in the `[watcher_datasources]` group and are named "
|
||||||
|
"`query_max_retries` and `query_timeout`."
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"Allow decision engine to pass strategy parameters, like optimization "
|
"Allow decision engine to pass strategy parameters, like optimization "
|
||||||
"threshold, to selected strategy, also strategy to provide parameters info to "
|
"threshold, to selected strategy, also strategy to provide parameters info to "
|
||||||
@@ -293,6 +454,34 @@ msgstr ""
|
|||||||
"threshold, to selected strategy, also strategy to provide parameters info to "
|
"threshold, to selected strategy, also strategy to provide parameters info to "
|
||||||
"end user."
|
"end user."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Allow using file to override metric map. Override the metric map of each "
|
||||||
|
"datasource as soon as it is created by the manager. This override comes from "
|
||||||
|
"a file whose path is provided by a setting in config file. The setting is "
|
||||||
|
"`watcher_decision_engine/metric_map_path`. The file contains a map per "
|
||||||
|
"datasource whose keys are the metric names as recognized by watcher and the "
|
||||||
|
"value is the real name of the metric in the datasource. This setting "
|
||||||
|
"defaults to `/etc/watcher/metric_map.yaml`, and presence of this file is "
|
||||||
|
"optional."
|
||||||
|
msgstr ""
|
||||||
|
"Allow using file to override metric map. Override the metric map of each "
|
||||||
|
"datasource as soon as it is created by the manager. This override comes from "
|
||||||
|
"a file whose path is provided by a setting in config file. The setting is "
|
||||||
|
"`watcher_decision_engine/metric_map_path`. The file contains a map per "
|
||||||
|
"datasource whose keys are the metric names as recognized by watcher and the "
|
||||||
|
"value is the real name of the metric in the datasource. This setting "
|
||||||
|
"defaults to `/etc/watcher/metric_map.yaml`, and presence of this file is "
|
||||||
|
"optional."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"An Watcher API WSGI application script ``watcher-api-wsgi`` is now "
|
||||||
|
"available. It is auto-generated by ``pbr`` and allows to run the API service "
|
||||||
|
"using WSGI server (for example Nginx and uWSGI)."
|
||||||
|
msgstr ""
|
||||||
|
"An Watcher API WSGI application script ``watcher-api-wsgi`` is now "
|
||||||
|
"available. It is auto-generated by ``pbr`` and allows to run the API service "
|
||||||
|
"using WSGI server (for example Nginx and uWSGI)."
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"Audits have 'name' field now, that is more friendly to end users. Audit's "
|
"Audits have 'name' field now, that is more friendly to end users. Audit's "
|
||||||
"name can't exceed 63 characters."
|
"name can't exceed 63 characters."
|
||||||
@@ -300,9 +489,25 @@ msgstr ""
|
|||||||
"Audits have 'name' field now, that is more friendly to end users. Audit's "
|
"Audits have 'name' field now, that is more friendly to end users. Audit's "
|
||||||
"name can't exceed 63 characters."
|
"name can't exceed 63 characters."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Baremetal Model gets Audit scoper with an ability to exclude Ironic nodes."
|
||||||
|
msgstr ""
|
||||||
|
"Baremetal Model gets Audit scope with an ability to exclude Ironic nodes."
|
||||||
|
|
||||||
msgid "Bug Fixes"
|
msgid "Bug Fixes"
|
||||||
msgstr "Bug Fixes"
|
msgstr "Bug Fixes"
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Ceilometer Datasource has been deprecated since its API has been deprecated "
|
||||||
|
"in Ocata cycle. Watcher has supported Ceilometer for some releases after "
|
||||||
|
"Ocata to let users migrate to Gnocchi/Monasca datasources. Since Train "
|
||||||
|
"release, Ceilometer support will be removed."
|
||||||
|
msgstr ""
|
||||||
|
"Ceilometer Datasource has been deprecated since its API has been deprecated "
|
||||||
|
"in Ocata cycle. Watcher has supported Ceilometer for some releases after "
|
||||||
|
"Ocata to let users migrate to Gnocchi/Monasca datasources. Since Train "
|
||||||
|
"release, Ceilometer support will be removed."
|
||||||
|
|
||||||
msgid "Centralize all configuration options for Watcher."
|
msgid "Centralize all configuration options for Watcher."
|
||||||
msgstr "Centralise all configuration options for Watcher."
|
msgstr "Centralise all configuration options for Watcher."
|
||||||
|
|
||||||
@@ -360,6 +565,52 @@ msgstr ""
|
|||||||
"Now instances from particular project in OpenStack can be excluded from "
|
"Now instances from particular project in OpenStack can be excluded from "
|
||||||
"audit defining scope in audit templates."
|
"audit defining scope in audit templates."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"For a large cloud infrastructure, retrieving data from Nova may take a long "
|
||||||
|
"time. To avoid getting too much data from Nova, building the compute data "
|
||||||
|
"model according to the scope of audit."
|
||||||
|
msgstr ""
|
||||||
|
"For a large cloud infrastructure, retrieving data from Nova may take a long "
|
||||||
|
"time. To avoid getting too much data from Nova, building the compute data "
|
||||||
|
"model according to the scope of audit."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Grafana has been added as datasource that can be used for collecting "
|
||||||
|
"metrics. The configuration options allow to specify what metrics and how "
|
||||||
|
"they are stored in grafana so that no matter how Grafana is configured it "
|
||||||
|
"can still be used. The configuration can be done via the typical "
|
||||||
|
"configuration file but it is recommended to configure most options in the "
|
||||||
|
"yaml file for metrics. For a complete walkthrough on configuring Grafana "
|
||||||
|
"see: https://docs.openstack.org/watcher/latest/datasources/grafana.html"
|
||||||
|
msgstr ""
|
||||||
|
"Grafana has been added as datasource that can be used for collecting "
|
||||||
|
"metrics. The configuration options allow to specify what metrics and how "
|
||||||
|
"they are stored in Grafana so that no matter how Grafana is configured it "
|
||||||
|
"can still be used. The configuration can be done via the typical "
|
||||||
|
"configuration file but it is recommended to configure most options in the "
|
||||||
|
"yaml file for metrics. For a complete walkthrough on configuring Grafana "
|
||||||
|
"see: https://docs.openstack.org/watcher/latest/datasources/grafana.html"
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"If Gnocchi was configured to have a custom amount of retries and or a custom "
|
||||||
|
"timeout then the configuration needs to moved into the "
|
||||||
|
"`[watcher_datasources]` group instead of the `[gnocchi_client]` group."
|
||||||
|
msgstr ""
|
||||||
|
"If Gnocchi was configured to have a custom amount of retries and or a custom "
|
||||||
|
"timeout then the configuration needs to moved into the "
|
||||||
|
"`[watcher_datasources]` group instead of the `[gnocchi_client]` group."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Improved interface for datasource baseclass that better defines expected "
|
||||||
|
"values and types for parameters and return types of all abstract methods. "
|
||||||
|
"This allows all strategies to work with every datasource provided the "
|
||||||
|
"metrics are configured for that given datasource."
|
||||||
|
msgstr ""
|
||||||
|
"Improved interface for datasource baseclass that better defines expected "
|
||||||
|
"values and types for parameters and return types of all abstract methods. "
|
||||||
|
"This allows all strategies to work with every datasource provided the "
|
||||||
|
"metrics are configured for that given datasource."
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"Instance cold migration logic is now replaced with using Nova migrate "
|
"Instance cold migration logic is now replaced with using Nova migrate "
|
||||||
"Server(migrate Action) API which has host option since v2.56."
|
"Server(migrate Action) API which has host option since v2.56."
|
||||||
@@ -367,6 +618,17 @@ msgstr ""
|
|||||||
"Instance cold migration logic is now replaced with using Nova migrate "
|
"Instance cold migration logic is now replaced with using Nova migrate "
|
||||||
"Server(migrate Action) API which has host option since v2.56."
|
"Server(migrate Action) API which has host option since v2.56."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Many operations in the decision engine will block on I/O. Such I/O "
|
||||||
|
"operations can stall the execution of a sequential application "
|
||||||
|
"significantly. To reduce the potential bottleneck of many operations the "
|
||||||
|
"general purpose decision engine threadpool is introduced."
|
||||||
|
msgstr ""
|
||||||
|
"Many operations in the decision engine will block on I/O. Such I/O "
|
||||||
|
"operations can stall the execution of a sequential application "
|
||||||
|
"significantly. To reduce the potential bottleneck of many operations the "
|
||||||
|
"general purpose decision engine threadpool is introduced."
|
||||||
|
|
||||||
msgid "New Features"
|
msgid "New Features"
|
||||||
msgstr "New Features"
|
msgstr "New Features"
|
||||||
|
|
||||||
@@ -389,6 +651,13 @@ msgstr ""
|
|||||||
"Nova API version is now set to 2.56 by default. This needs the migrate "
|
"Nova API version is now set to 2.56 by default. This needs the migrate "
|
||||||
"action of migration type cold with destination_node parameter to work."
|
"action of migration type cold with destination_node parameter to work."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Now Watcher strategy can select specific planner beyond default. Strategy "
|
||||||
|
"can set planner property to specify its own planner."
|
||||||
|
msgstr ""
|
||||||
|
"Now Watcher strategy can select specific planner beyond default. Strategy "
|
||||||
|
"can set planner property to specify its own planner."
|
||||||
|
|
||||||
msgid "Ocata Series Release Notes"
|
msgid "Ocata Series Release Notes"
|
||||||
msgstr "Ocata Series Release Notes"
|
msgstr "Ocata Series Release Notes"
|
||||||
|
|
||||||
@@ -429,12 +698,77 @@ msgstr ""
|
|||||||
"resources will be called \"Audit scope\" and will be defined in each audit "
|
"resources will be called \"Audit scope\" and will be defined in each audit "
|
||||||
"template (which contains the audit settings)."
|
"template (which contains the audit settings)."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Python 2.7 support has been dropped. Last release of Watcher to support "
|
||||||
|
"py2.7 is OpenStack Train. The minimum version of Python now supported by "
|
||||||
|
"Watcher is Python 3.6."
|
||||||
|
msgstr ""
|
||||||
|
"Python 2.7 support has been dropped. Last release of Watcher to support "
|
||||||
|
"py2.7 is OpenStack Train. The minimum version of Python now supported by "
|
||||||
|
"Watcher is Python 3.6."
|
||||||
|
|
||||||
msgid "Queens Series Release Notes"
|
msgid "Queens Series Release Notes"
|
||||||
msgstr "Queens Series Release Notes"
|
msgstr "Queens Series Release Notes"
|
||||||
|
|
||||||
msgid "Rocky Series Release Notes"
|
msgid "Rocky Series Release Notes"
|
||||||
msgstr "Rocky Series Release Notes"
|
msgstr "Rocky Series Release Notes"
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Several strategies have changed the `node` parameter to `compute_node` to be "
|
||||||
|
"better aligned with terminology. These strategies include "
|
||||||
|
"`basic_consolidation` and `workload_stabilzation`. The `node` parameter will "
|
||||||
|
"remain supported during Train release and will be removed in the subsequent "
|
||||||
|
"release."
|
||||||
|
msgstr ""
|
||||||
|
"Several strategies have changed the `node` parameter to `compute_node` to be "
|
||||||
|
"better aligned with terminology. These strategies include "
|
||||||
|
"`basic_consolidation` and `workload_stabilzation`. The `node` parameter will "
|
||||||
|
"remain supported during Train release and will be removed in the subsequent "
|
||||||
|
"release."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Specific strategies can override this order and use datasources which are "
|
||||||
|
"not listed in the global preference."
|
||||||
|
msgstr ""
|
||||||
|
"Specific strategies can override this order and use datasources which are "
|
||||||
|
"not listed in the global preference."
|
||||||
|
|
||||||
|
msgid "Stein Series Release Notes"
|
||||||
|
msgstr "Stein Series Release Notes"
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"The building of the compute (Nova) data model will be done using the "
|
||||||
|
"decision engine threadpool, thereby, significantly reducing the total time "
|
||||||
|
"required to build it."
|
||||||
|
msgstr ""
|
||||||
|
"The building of the compute (Nova) data model will be done using the "
|
||||||
|
"decision engine threadpool, thereby, significantly reducing the total time "
|
||||||
|
"required to build it."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"The configuration options for query retries in `[gnocchi_client]` are "
|
||||||
|
"deprecated and the option in `[watcher_datasources]` should now be used."
|
||||||
|
msgstr ""
|
||||||
|
"The configuration options for query retries in `[gnocchi_client]` are "
|
||||||
|
"deprecated and the option in `[watcher_datasources]` should now be used."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"The default value of ``[oslo_policy] policy_file`` config option has been "
|
||||||
|
"changed from ``policy.json`` to ``policy.yaml``. Operators who are utilizing "
|
||||||
|
"customized or previously generated static policy JSON files (which are not "
|
||||||
|
"needed by default), should generate new policy files or convert them in YAML "
|
||||||
|
"format. Use the `oslopolicy-convert-json-to-yaml <https://docs.openstack.org/"
|
||||||
|
"oslo.policy/latest/cli/oslopolicy-convert-json-to-yaml.html>`_ tool to "
|
||||||
|
"convert a JSON to YAML formatted policy file in backward compatible way."
|
||||||
|
msgstr ""
|
||||||
|
"The default value of ``[oslo_policy] policy_file`` config option has been "
|
||||||
|
"changed from ``policy.json`` to ``policy.yaml``. Operators who are utilizing "
|
||||||
|
"customized or previously generated static policy JSON files (which are not "
|
||||||
|
"needed by default), should generate new policy files or convert them in YAML "
|
||||||
|
"format. Use the `oslopolicy-convert-json-to-yaml <https://docs.openstack.org/"
|
||||||
|
"oslo.policy/latest/cli/oslopolicy-convert-json-to-yaml.html>`_ tool to "
|
||||||
|
"convert a JSON to YAML formatted policy file in backward compatible way."
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"The graph model describes how VMs are associated to compute hosts. This "
|
"The graph model describes how VMs are associated to compute hosts. This "
|
||||||
"allows for seeing relationships upfront between the entities and hence can "
|
"allows for seeing relationships upfront between the entities and hence can "
|
||||||
@@ -455,6 +789,22 @@ msgstr ""
|
|||||||
"was fixed. Before fixing, it booted an instance in the service project as a "
|
"was fixed. Before fixing, it booted an instance in the service project as a "
|
||||||
"migrated instance."
|
"migrated instance."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"The minimum required version of the ``[nova_client]/api_version`` value is "
|
||||||
|
"now enforced to be ``2.56`` which is available since the Queens version of "
|
||||||
|
"the nova compute service."
|
||||||
|
msgstr ""
|
||||||
|
"The minimum required version of the ``[nova_client]/api_version`` value is "
|
||||||
|
"now enforced to be ``2.56`` which is available since the Queens version of "
|
||||||
|
"the Nova compute service."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"The new strategy baseclass has significant changes in method parameters and "
|
||||||
|
"any out-of-tree strategies will have to be adopted."
|
||||||
|
msgstr ""
|
||||||
|
"The new strategy baseclass has significant changes in method parameters and "
|
||||||
|
"any out-of-tree strategies will have to be adopted."
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"There is new ability to create Watcher continuous audits with cron interval. "
|
"There is new ability to create Watcher continuous audits with cron interval. "
|
||||||
"It means you may use, for example, optional argument '--interval \"\\*/5 \\* "
|
"It means you may use, for example, optional argument '--interval \"\\*/5 \\* "
|
||||||
@@ -468,9 +818,45 @@ msgstr ""
|
|||||||
"best effort basis and therefore, we recommend you to use a minimal cron "
|
"best effort basis and therefore, we recommend you to use a minimal cron "
|
||||||
"interval of at least one minute."
|
"interval of at least one minute."
|
||||||
|
|
||||||
|
msgid "Train Series Release Notes"
|
||||||
|
msgstr "Train Series Release Notes"
|
||||||
|
|
||||||
msgid "Upgrade Notes"
|
msgid "Upgrade Notes"
|
||||||
msgstr "Upgrade Notes"
|
msgstr "Upgrade Notes"
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Use of JSON policy files was deprecated by the ``oslo.policy`` library "
|
||||||
|
"during the Victoria development cycle. As a result, this deprecation is "
|
||||||
|
"being noted in the Wallaby cycle with an anticipated future removal of "
|
||||||
|
"support by ``oslo.policy``. As such operators will need to convert to YAML "
|
||||||
|
"policy files. Please see the upgrade notes for details on migration of any "
|
||||||
|
"custom policy files."
|
||||||
|
msgstr ""
|
||||||
|
"Use of JSON policy files was deprecated by the ``oslo.policy`` library "
|
||||||
|
"during the Victoria development cycle. As a result, this deprecation is "
|
||||||
|
"being noted in the Wallaby cycle with an anticipated future removal of "
|
||||||
|
"support by ``oslo.policy``. As such operators will need to convert to YAML "
|
||||||
|
"policy files. Please see the upgrade notes for details on migration of any "
|
||||||
|
"custom policy files."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Using ``watcher/api/app.wsgi`` script is deprecated and it will be removed "
|
||||||
|
"in U release. Please switch to automatically generated ``watcher-api-wsgi`` "
|
||||||
|
"script instead."
|
||||||
|
msgstr ""
|
||||||
|
"Using ``watcher/api/app.wsgi`` script is deprecated and it will be removed "
|
||||||
|
"in U release. Please switch to automatically generated ``watcher-api-wsgi`` "
|
||||||
|
"script instead."
|
||||||
|
|
||||||
|
msgid "Ussuri Series Release Notes"
|
||||||
|
msgstr "Ussuri Series Release Notes"
|
||||||
|
|
||||||
|
msgid "Victoria Series Release Notes"
|
||||||
|
msgstr "Victoria Series Release Notes"
|
||||||
|
|
||||||
|
msgid "Wallaby Series Release Notes"
|
||||||
|
msgstr "Wallaby Series Release Notes"
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"Watcher can continuously optimize the OpenStack cloud for a specific "
|
"Watcher can continuously optimize the OpenStack cloud for a specific "
|
||||||
"strategy or goal by triggering an audit periodically which generates an "
|
"strategy or goal by triggering an audit periodically which generates an "
|
||||||
@@ -480,6 +866,15 @@ msgstr ""
|
|||||||
"strategy or goal by triggering an audit periodically which generates an "
|
"strategy or goal by triggering an audit periodically which generates an "
|
||||||
"action plan and run it automatically."
|
"action plan and run it automatically."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Watcher can get resource information such as total, allocation ratio and "
|
||||||
|
"reserved information from Placement API. Now we add some new fields to the "
|
||||||
|
"Watcher Data Model:"
|
||||||
|
msgstr ""
|
||||||
|
"Watcher can get resource information such as total, allocation ratio and "
|
||||||
|
"reserved information from Placement API. Now we add some new fields to the "
|
||||||
|
"Watcher Data Model:"
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"Watcher can now run specific actions in parallel improving the performances "
|
"Watcher can now run specific actions in parallel improving the performances "
|
||||||
"dramatically when executing an action plan."
|
"dramatically when executing an action plan."
|
||||||
@@ -517,6 +912,15 @@ msgstr ""
|
|||||||
"includes all instances. It filters excluded instances when migration during "
|
"includes all instances. It filters excluded instances when migration during "
|
||||||
"the audit."
|
"the audit."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"Watcher now supports configuring which datasource to use and in which order. "
|
||||||
|
"This configuration is done by specifying datasources in the "
|
||||||
|
"watcher_datasources section:"
|
||||||
|
msgstr ""
|
||||||
|
"Watcher now supports configuring which datasource to use and in which order. "
|
||||||
|
"This configuration is done by specifying datasources in the "
|
||||||
|
"watcher_datasources section:"
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"Watcher removes the support to Nova legacy notifications because of Nova "
|
"Watcher removes the support to Nova legacy notifications because of Nova "
|
||||||
"will deprecate them."
|
"will deprecate them."
|
||||||
@@ -557,9 +961,24 @@ msgstr ""
|
|||||||
"Watcher supports multiple metrics backend and relies on Ceilometer and "
|
"Watcher supports multiple metrics backend and relies on Ceilometer and "
|
||||||
"Monasca."
|
"Monasca."
|
||||||
|
|
||||||
|
msgid "We also add some new propeties:"
|
||||||
|
msgstr "We also add some new properties:"
|
||||||
|
|
||||||
msgid "Welcome to watcher's Release Notes documentation!"
|
msgid "Welcome to watcher's Release Notes documentation!"
|
||||||
msgstr "Welcome to watcher's Release Notes documentation!"
|
msgstr "Welcome to watcher's Release Notes documentation!"
|
||||||
|
|
||||||
|
msgid "Xena Series Release Notes"
|
||||||
|
msgstr "Xena Series Release Notes"
|
||||||
|
|
||||||
|
msgid "Yoga Series Release Notes"
|
||||||
|
msgstr "Yoga Series Release Notes"
|
||||||
|
|
||||||
|
msgid "Zed Series Release Notes"
|
||||||
|
msgstr "Zed Series Release Notes"
|
||||||
|
|
||||||
|
msgid "``[watcher_datasources] datasources = gnocchi,monasca,ceilometer``"
|
||||||
|
msgstr "``[watcher_datasources] datasources = gnocchi,monasca,ceilometer``"
|
||||||
|
|
||||||
msgid ""
|
msgid ""
|
||||||
"all Watcher objects have been refactored to support OVO (oslo."
|
"all Watcher objects have been refactored to support OVO (oslo."
|
||||||
"versionedobjects) which was a prerequisite step in order to implement "
|
"versionedobjects) which was a prerequisite step in order to implement "
|
||||||
@@ -569,6 +988,21 @@ msgstr ""
|
|||||||
"versionedobjects) which was a prerequisite step in order to implement "
|
"versionedobjects) which was a prerequisite step in order to implement "
|
||||||
"versioned notifications."
|
"versioned notifications."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"disk_gb_capacity: The amount of disk, take allocation ratio into account, "
|
||||||
|
"but do not include reserved."
|
||||||
|
msgstr ""
|
||||||
|
"disk_gb_capacity: The amount of disk, take allocation ratio into account, "
|
||||||
|
"but do not include reserved."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"disk_gb_reserved: The amount of disk a node has reserved for its own use."
|
||||||
|
msgstr ""
|
||||||
|
"disk_gb_reserved: The amount of disk a node has reserved for its own use."
|
||||||
|
|
||||||
|
msgid "disk_ratio: Disk allocation ratio."
|
||||||
|
msgstr "disk_ratio: Disk allocation ratio."
|
||||||
|
|
||||||
msgid "instance.create.end"
|
msgid "instance.create.end"
|
||||||
msgstr "instance.create.end"
|
msgstr "instance.create.end"
|
||||||
|
|
||||||
@@ -635,6 +1069,21 @@ msgstr "instance.unshelve.end"
|
|||||||
msgid "instance.update"
|
msgid "instance.update"
|
||||||
msgstr "instance.update"
|
msgstr "instance.update"
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"memory_mb_capacity: The amount of memory, take allocation ratio into "
|
||||||
|
"account, but do not include reserved."
|
||||||
|
msgstr ""
|
||||||
|
"memory_mb_capacity: The amount of memory, take allocation ratio into "
|
||||||
|
"account, but do not include reserved."
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"memory_mb_reserved: The amount of memory a node has reserved for its own use."
|
||||||
|
msgstr ""
|
||||||
|
"memory_mb_reserved: The amount of memory a node has reserved for its own use."
|
||||||
|
|
||||||
|
msgid "memory_ratio: Memory allocation ratio."
|
||||||
|
msgstr "memory_ratio: Memory allocation ratio."
|
||||||
|
|
||||||
msgid "new:"
|
msgid "new:"
|
||||||
msgstr "new:"
|
msgstr "new:"
|
||||||
|
|
||||||
@@ -649,3 +1098,16 @@ msgstr "service.delete"
|
|||||||
|
|
||||||
msgid "service.update"
|
msgid "service.update"
|
||||||
msgstr "service.update"
|
msgstr "service.update"
|
||||||
|
|
||||||
|
msgid ""
|
||||||
|
"vcpu_capacity: The amount of vcpu, take allocation ratio into account, but "
|
||||||
|
"do not include reserved."
|
||||||
|
msgstr ""
|
||||||
|
"vcpu_capacity: The amount of vcpu, take allocation ratio into account, but "
|
||||||
|
"do not include reserved."
|
||||||
|
|
||||||
|
msgid "vcpu_ratio: CPU allocation ratio."
|
||||||
|
msgstr "vcpu_ratio: CPU allocation ratio."
|
||||||
|
|
||||||
|
msgid "vcpu_reserved: The amount of cpu a node has reserved for its own use."
|
||||||
|
msgstr "vcpu_reserved: The amount of CPU a node has reserved for its own use."
|
||||||
|
|||||||
6
releasenotes/source/ussuri.rst
Normal file
6
releasenotes/source/ussuri.rst
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
===========================
|
||||||
|
Ussuri Series Release Notes
|
||||||
|
===========================
|
||||||
|
|
||||||
|
.. release-notes::
|
||||||
|
:branch: stable/ussuri
|
||||||
6
releasenotes/source/victoria.rst
Normal file
6
releasenotes/source/victoria.rst
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
=============================
|
||||||
|
Victoria Series Release Notes
|
||||||
|
=============================
|
||||||
|
|
||||||
|
.. release-notes::
|
||||||
|
:branch: stable/victoria
|
||||||
6
releasenotes/source/wallaby.rst
Normal file
6
releasenotes/source/wallaby.rst
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
============================
|
||||||
|
Wallaby Series Release Notes
|
||||||
|
============================
|
||||||
|
|
||||||
|
.. release-notes::
|
||||||
|
:branch: stable/wallaby
|
||||||
6
releasenotes/source/xena.rst
Normal file
6
releasenotes/source/xena.rst
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
=========================
|
||||||
|
Xena Series Release Notes
|
||||||
|
=========================
|
||||||
|
|
||||||
|
.. release-notes::
|
||||||
|
:branch: stable/xena
|
||||||
6
releasenotes/source/yoga.rst
Normal file
6
releasenotes/source/yoga.rst
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
=========================
|
||||||
|
Yoga Series Release Notes
|
||||||
|
=========================
|
||||||
|
|
||||||
|
.. release-notes::
|
||||||
|
:branch: stable/yoga
|
||||||
6
releasenotes/source/zed.rst
Normal file
6
releasenotes/source/zed.rst
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
========================
|
||||||
|
Zed Series Release Notes
|
||||||
|
========================
|
||||||
|
|
||||||
|
.. release-notes::
|
||||||
|
:branch: stable/zed
|
||||||
@@ -1,36 +1,35 @@
|
|||||||
# The order of packages is significant, because pip processes them in the order
|
# The order of packages is significant, because pip processes them in the order
|
||||||
# of appearance. Changing the order has an impact on the overall integration
|
# of appearance. Changing the order has an impact on the overall integration
|
||||||
# process, which may cause wedges in the gate later.
|
# process, which may cause wedges in the gate later.
|
||||||
|
|
||||||
apscheduler>=3.5.1 # MIT License
|
apscheduler>=3.5.1 # MIT License
|
||||||
jsonpatch>=1.21 # BSD
|
jsonpatch>=1.21 # BSD
|
||||||
keystoneauth1>=3.4.0 # Apache-2.0
|
keystoneauth1>=3.4.0 # Apache-2.0
|
||||||
jsonschema>=2.6.0 # MIT
|
jsonschema>=3.2.0 # MIT
|
||||||
keystonemiddleware>=4.21.0 # Apache-2.0
|
keystonemiddleware>=4.21.0 # Apache-2.0
|
||||||
lxml>=4.1.1 # BSD
|
lxml>=4.5.1 # BSD
|
||||||
croniter>=0.3.20 # MIT License
|
croniter>=0.3.20 # MIT License
|
||||||
os-resource-classes>=0.4.0
|
os-resource-classes>=0.4.0
|
||||||
oslo.concurrency>=3.26.0 # Apache-2.0
|
oslo.concurrency>=3.26.0 # Apache-2.0
|
||||||
oslo.cache>=1.29.0 # Apache-2.0
|
oslo.cache>=1.29.0 # Apache-2.0
|
||||||
oslo.config>=5.2.0 # Apache-2.0
|
oslo.config>=6.8.0 # Apache-2.0
|
||||||
oslo.context>=2.21.0 # Apache-2.0
|
oslo.context>=2.21.0 # Apache-2.0
|
||||||
oslo.db>=4.35.0 # Apache-2.0
|
oslo.db>=4.44.0 # Apache-2.0
|
||||||
oslo.i18n>=3.20.0 # Apache-2.0
|
oslo.i18n>=3.20.0 # Apache-2.0
|
||||||
oslo.log>=3.37.0 # Apache-2.0
|
oslo.log>=3.37.0 # Apache-2.0
|
||||||
oslo.messaging>=8.1.2 # Apache-2.0
|
oslo.messaging>=14.1.0 # Apache-2.0
|
||||||
oslo.policy>=1.34.0 # Apache-2.0
|
oslo.policy>=3.6.0 # Apache-2.0
|
||||||
oslo.reports>=1.27.0 # Apache-2.0
|
oslo.reports>=1.27.0 # Apache-2.0
|
||||||
oslo.serialization>=2.25.0 # Apache-2.0
|
oslo.serialization>=2.25.0 # Apache-2.0
|
||||||
oslo.service>=1.30.0 # Apache-2.0
|
oslo.service>=1.30.0 # Apache-2.0
|
||||||
oslo.upgradecheck>=0.1.0 # Apache-2.0
|
oslo.upgradecheck>=1.3.0 # Apache-2.0
|
||||||
oslo.utils>=3.36.0 # Apache-2.0
|
oslo.utils>=3.36.0 # Apache-2.0
|
||||||
oslo.versionedobjects>=1.32.0 # Apache-2.0
|
oslo.versionedobjects>=1.32.0 # Apache-2.0
|
||||||
PasteDeploy>=1.5.2 # MIT
|
PasteDeploy>=1.5.2 # MIT
|
||||||
pbr>=3.1.1 # Apache-2.0
|
pbr>=3.1.1 # Apache-2.0
|
||||||
pecan>=1.3.2 # BSD
|
pecan>=1.3.2 # BSD
|
||||||
PrettyTable<0.8,>=0.7.2 # BSD
|
PrettyTable>=0.7.2 # BSD
|
||||||
gnocchiclient>=7.0.1 # Apache-2.0
|
gnocchiclient>=7.0.1 # Apache-2.0
|
||||||
python-ceilometerclient>=2.9.0 # Apache-2.0
|
|
||||||
python-cinderclient>=3.5.0 # Apache-2.0
|
python-cinderclient>=3.5.0 # Apache-2.0
|
||||||
python-glanceclient>=2.9.1 # Apache-2.0
|
python-glanceclient>=2.9.1 # Apache-2.0
|
||||||
python-keystoneclient>=3.15.0 # Apache-2.0
|
python-keystoneclient>=3.15.0 # Apache-2.0
|
||||||
@@ -39,12 +38,11 @@ python-neutronclient>=6.7.0 # Apache-2.0
|
|||||||
python-novaclient>=14.1.0 # Apache-2.0
|
python-novaclient>=14.1.0 # Apache-2.0
|
||||||
python-openstackclient>=3.14.0 # Apache-2.0
|
python-openstackclient>=3.14.0 # Apache-2.0
|
||||||
python-ironicclient>=2.5.0 # Apache-2.0
|
python-ironicclient>=2.5.0 # Apache-2.0
|
||||||
six>=1.11.0 # MIT
|
|
||||||
SQLAlchemy>=1.2.5 # MIT
|
SQLAlchemy>=1.2.5 # MIT
|
||||||
stevedore>=1.28.0 # Apache-2.0
|
stevedore>=1.28.0 # Apache-2.0
|
||||||
taskflow>=3.7.1 # Apache-2.0
|
taskflow>=3.8.0 # Apache-2.0
|
||||||
WebOb>=1.8.5 # MIT
|
WebOb>=1.8.5 # MIT
|
||||||
WSME>=0.9.2 # MIT
|
WSME>=0.9.2 # MIT
|
||||||
networkx>=2.2;python_version>='3.4' # BSD
|
networkx>=2.4 # BSD
|
||||||
microversion_parse>=0.2.1 # Apache-2.0
|
microversion_parse>=0.2.1 # Apache-2.0
|
||||||
futurist>=1.8.0 # Apache-2.0
|
futurist>=1.8.0 # Apache-2.0
|
||||||
|
|||||||
29
setup.cfg
29
setup.cfg
@@ -1,12 +1,12 @@
|
|||||||
[metadata]
|
[metadata]
|
||||||
name = python-watcher
|
name = python-watcher
|
||||||
summary = OpenStack Watcher provides a flexible and scalable resource optimization service for multi-tenant OpenStack-based clouds.
|
summary = OpenStack Watcher provides a flexible and scalable resource optimization service for multi-tenant OpenStack-based clouds.
|
||||||
description-file =
|
description_file =
|
||||||
README.rst
|
README.rst
|
||||||
author = OpenStack
|
author = OpenStack
|
||||||
author-email = openstack-discuss@lists.openstack.org
|
author_email = openstack-discuss@lists.openstack.org
|
||||||
home-page = https://docs.openstack.org/watcher/latest/
|
home_page = https://docs.openstack.org/watcher/latest/
|
||||||
python-requires = >=3.6
|
python_requires = >=3.8
|
||||||
classifier =
|
classifier =
|
||||||
Environment :: OpenStack
|
Environment :: OpenStack
|
||||||
Intended Audience :: Information Technology
|
Intended Audience :: Information Technology
|
||||||
@@ -17,8 +17,10 @@ classifier =
|
|||||||
Programming Language :: Python :: Implementation :: CPython
|
Programming Language :: Python :: Implementation :: CPython
|
||||||
Programming Language :: Python :: 3 :: Only
|
Programming Language :: Python :: 3 :: Only
|
||||||
Programming Language :: Python :: 3
|
Programming Language :: Python :: 3
|
||||||
Programming Language :: Python :: 3.6
|
Programming Language :: Python :: 3.8
|
||||||
Programming Language :: Python :: 3.7
|
Programming Language :: Python :: 3.9
|
||||||
|
Programming Language :: Python :: 3.10
|
||||||
|
Programming Language :: Python :: 3.11
|
||||||
|
|
||||||
[files]
|
[files]
|
||||||
packages =
|
packages =
|
||||||
@@ -107,18 +109,3 @@ watcher_cluster_data_model_collectors =
|
|||||||
compute = watcher.decision_engine.model.collector.nova:NovaClusterDataModelCollector
|
compute = watcher.decision_engine.model.collector.nova:NovaClusterDataModelCollector
|
||||||
storage = watcher.decision_engine.model.collector.cinder:CinderClusterDataModelCollector
|
storage = watcher.decision_engine.model.collector.cinder:CinderClusterDataModelCollector
|
||||||
baremetal = watcher.decision_engine.model.collector.ironic:BaremetalClusterDataModelCollector
|
baremetal = watcher.decision_engine.model.collector.ironic:BaremetalClusterDataModelCollector
|
||||||
|
|
||||||
|
|
||||||
[compile_catalog]
|
|
||||||
directory = watcher/locale
|
|
||||||
domain = watcher
|
|
||||||
|
|
||||||
[update_catalog]
|
|
||||||
domain = watcher
|
|
||||||
output_dir = watcher/locale
|
|
||||||
input_file = watcher/locale/watcher.pot
|
|
||||||
|
|
||||||
[extract_messages]
|
|
||||||
keywords = _ gettext ngettext l_ lazy_gettext _LI _LW _LE _LC
|
|
||||||
mapping_file = babel.cfg
|
|
||||||
output_file = watcher/locale/watcher.pot
|
|
||||||
|
|||||||
@@ -5,12 +5,11 @@
|
|||||||
coverage>=4.5.1 # Apache-2.0
|
coverage>=4.5.1 # Apache-2.0
|
||||||
doc8>=0.8.0 # Apache-2.0
|
doc8>=0.8.0 # Apache-2.0
|
||||||
freezegun>=0.3.10 # Apache-2.0
|
freezegun>=0.3.10 # Apache-2.0
|
||||||
hacking>=3.0,<3.1.0 # Apache-2.0
|
hacking>=3.0.1,<3.1.0 # Apache-2.0
|
||||||
mock>=2.0.0 # BSD
|
|
||||||
oslotest>=3.3.0 # Apache-2.0
|
oslotest>=3.3.0 # Apache-2.0
|
||||||
os-testr>=1.0.0 # Apache-2.0
|
|
||||||
testscenarios>=0.5.0 # Apache-2.0/BSD
|
testscenarios>=0.5.0 # Apache-2.0/BSD
|
||||||
testtools>=2.3.0 # MIT
|
testtools>=2.3.0 # MIT
|
||||||
stestr>=2.0.0 # Apache-2.0
|
stestr>=2.0.0 # Apache-2.0
|
||||||
os-api-ref>=1.4.0 # Apache-2.0
|
os-api-ref>=1.4.0 # Apache-2.0
|
||||||
bandit>=1.6.0 # Apache-2.0
|
bandit>=1.6.0 # Apache-2.0
|
||||||
|
WebTest>=2.0.27 # MIT
|
||||||
79
tox.ini
79
tox.ini
@@ -1,37 +1,41 @@
|
|||||||
[tox]
|
[tox]
|
||||||
minversion = 2.0
|
minversion = 3.18.0
|
||||||
envlist = py36,py37,pep8
|
envlist = py3,pep8
|
||||||
skipsdist = True
|
|
||||||
ignore_basepython_conflict = True
|
ignore_basepython_conflict = True
|
||||||
|
|
||||||
[testenv]
|
[testenv]
|
||||||
basepython = python3
|
basepython = python3
|
||||||
usedevelop = True
|
usedevelop = True
|
||||||
whitelist_externals = find
|
allowlist_externals = find
|
||||||
rm
|
rm
|
||||||
install_command = pip install {opts} {packages}
|
install_command = pip install -c{env:TOX_CONSTRAINTS_FILE:https://releases.openstack.org/constraints/upper/master} {opts} {packages}
|
||||||
setenv =
|
setenv =
|
||||||
VIRTUAL_ENV={envdir}
|
VIRTUAL_ENV={envdir}
|
||||||
deps =
|
deps =
|
||||||
-c{env:UPPER_CONSTRAINTS_FILE:https://releases.openstack.org/constraints/upper/master}
|
|
||||||
-r{toxinidir}/test-requirements.txt
|
-r{toxinidir}/test-requirements.txt
|
||||||
-r{toxinidir}/requirements.txt
|
-r{toxinidir}/requirements.txt
|
||||||
|
python-libmaas>=0.6.8
|
||||||
commands =
|
commands =
|
||||||
rm -f .testrepository/times.dbm
|
rm -f .testrepository/times.dbm
|
||||||
find . -type f -name "*.py[c|o]" -delete
|
find . -type f -name "*.py[c|o]" -delete
|
||||||
stestr run {posargs}
|
stestr run {posargs}
|
||||||
passenv = http_proxy HTTP_PROXY https_proxy HTTPS_PROXY no_proxy NO_PROXY
|
passenv =
|
||||||
|
http_proxy
|
||||||
|
HTTP_PROXY
|
||||||
|
https_proxy
|
||||||
|
HTTPS_PROXY
|
||||||
|
no_proxy
|
||||||
|
NO_PROXY
|
||||||
|
|
||||||
[testenv:pep8]
|
[testenv:pep8]
|
||||||
commands =
|
commands =
|
||||||
doc8 doc/source/ CONTRIBUTING.rst HACKING.rst README.rst
|
doc8 doc/source/ CONTRIBUTING.rst HACKING.rst README.rst
|
||||||
flake8
|
flake8
|
||||||
bandit -r watcher -x watcher/tests/* -n5 -ll -s B320
|
#bandit -r watcher -x watcher/tests/* -n5 -ll -s B320
|
||||||
|
|
||||||
[testenv:venv]
|
[testenv:venv]
|
||||||
setenv = PYTHONHASHSEED=0
|
setenv = PYTHONHASHSEED=0
|
||||||
deps =
|
deps =
|
||||||
-c{env:UPPER_CONSTRAINTS_FILE:https://releases.openstack.org/constraints/upper/master}
|
|
||||||
-r{toxinidir}/doc/requirements.txt
|
-r{toxinidir}/doc/requirements.txt
|
||||||
-r{toxinidir}/test-requirements.txt
|
-r{toxinidir}/test-requirements.txt
|
||||||
-r{toxinidir}/requirements.txt
|
-r{toxinidir}/requirements.txt
|
||||||
@@ -49,14 +53,15 @@ commands =
|
|||||||
|
|
||||||
[testenv:docs]
|
[testenv:docs]
|
||||||
setenv = PYTHONHASHSEED=0
|
setenv = PYTHONHASHSEED=0
|
||||||
deps = -r{toxinidir}/doc/requirements.txt
|
deps =
|
||||||
|
-r{toxinidir}/doc/requirements.txt
|
||||||
commands =
|
commands =
|
||||||
rm -fr doc/build doc/source/api/ .autogenerated
|
rm -fr doc/build doc/source/api/ .autogenerated
|
||||||
sphinx-build -W --keep-going -b html doc/source doc/build/html
|
sphinx-build -W --keep-going -b html doc/source doc/build/html
|
||||||
|
|
||||||
[testenv:api-ref]
|
[testenv:api-ref]
|
||||||
deps = -r{toxinidir}/doc/requirements.txt
|
deps = -r{toxinidir}/doc/requirements.txt
|
||||||
whitelist_externals = bash
|
allowlist_externals = bash
|
||||||
commands =
|
commands =
|
||||||
bash -c 'rm -rf api-ref/build'
|
bash -c 'rm -rf api-ref/build'
|
||||||
sphinx-build -W --keep-going -b html -d api-ref/build/doctrees api-ref/source api-ref/build/html
|
sphinx-build -W --keep-going -b html -d api-ref/build/doctrees api-ref/source api-ref/build/html
|
||||||
@@ -73,6 +78,28 @@ commands =
|
|||||||
commands =
|
commands =
|
||||||
oslopolicy-sample-generator --config-file etc/watcher/oslo-policy-generator/watcher-policy-generator.conf
|
oslopolicy-sample-generator --config-file etc/watcher/oslo-policy-generator/watcher-policy-generator.conf
|
||||||
|
|
||||||
|
[testenv:wheel]
|
||||||
|
commands = python setup.py bdist_wheel
|
||||||
|
|
||||||
|
[testenv:pdf-docs]
|
||||||
|
envdir = {toxworkdir}/docs
|
||||||
|
deps = {[testenv:docs]deps}
|
||||||
|
allowlist_externals =
|
||||||
|
rm
|
||||||
|
make
|
||||||
|
commands =
|
||||||
|
rm -rf doc/build/pdf
|
||||||
|
sphinx-build -W --keep-going -b latex doc/source doc/build/pdf
|
||||||
|
make -C doc/build/pdf
|
||||||
|
|
||||||
|
[testenv:releasenotes]
|
||||||
|
deps = -r{toxinidir}/doc/requirements.txt
|
||||||
|
commands = sphinx-build -a -W -E -d releasenotes/build/doctrees --keep-going -b html releasenotes/source releasenotes/build/html
|
||||||
|
|
||||||
|
[testenv:bandit]
|
||||||
|
deps = -r{toxinidir}/test-requirements.txt
|
||||||
|
commands = bandit -r watcher -x watcher/tests/* -n5 -ll -s B320
|
||||||
|
|
||||||
[flake8]
|
[flake8]
|
||||||
filename = *.py,app.wsgi
|
filename = *.py,app.wsgi
|
||||||
show-source=True
|
show-source=True
|
||||||
@@ -82,9 +109,6 @@ builtins= _
|
|||||||
enable-extensions = H106,H203,H904
|
enable-extensions = H106,H203,H904
|
||||||
exclude=.venv,.git,.tox,dist,doc,*lib/python*,*egg,build,*sqlalchemy/alembic/versions/*,demo/,releasenotes
|
exclude=.venv,.git,.tox,dist,doc,*lib/python*,*egg,build,*sqlalchemy/alembic/versions/*,demo/,releasenotes
|
||||||
|
|
||||||
[testenv:wheel]
|
|
||||||
commands = python setup.py bdist_wheel
|
|
||||||
|
|
||||||
[hacking]
|
[hacking]
|
||||||
import_exceptions = watcher._i18n
|
import_exceptions = watcher._i18n
|
||||||
|
|
||||||
@@ -105,35 +129,10 @@ extension =
|
|||||||
N340 = checks:check_oslo_i18n_wrapper
|
N340 = checks:check_oslo_i18n_wrapper
|
||||||
N341 = checks:check_builtins_gettext
|
N341 = checks:check_builtins_gettext
|
||||||
N342 = checks:no_redundant_import_alias
|
N342 = checks:no_redundant_import_alias
|
||||||
|
N366 = checks:import_stock_mock
|
||||||
paths = ./watcher/hacking
|
paths = ./watcher/hacking
|
||||||
|
|
||||||
|
|
||||||
[doc8]
|
[doc8]
|
||||||
extension=.rst
|
extension=.rst
|
||||||
# todo: stop ignoring doc/source/man when https://bugs.launchpad.net/doc8/+bug/1502391 is fixed
|
# todo: stop ignoring doc/source/man when https://bugs.launchpad.net/doc8/+bug/1502391 is fixed
|
||||||
ignore-path=doc/source/image_src,doc/source/man,doc/source/api
|
ignore-path=doc/source/image_src,doc/source/man,doc/source/api
|
||||||
|
|
||||||
[testenv:pdf-docs]
|
|
||||||
envdir = {toxworkdir}/docs
|
|
||||||
deps = {[testenv:docs]deps}
|
|
||||||
whitelist_externals =
|
|
||||||
rm
|
|
||||||
make
|
|
||||||
commands =
|
|
||||||
rm -rf doc/build/pdf
|
|
||||||
sphinx-build -W --keep-going -b latex doc/source doc/build/pdf
|
|
||||||
make -C doc/build/pdf
|
|
||||||
|
|
||||||
[testenv:releasenotes]
|
|
||||||
deps = -r{toxinidir}/doc/requirements.txt
|
|
||||||
commands = sphinx-build -a -W -E -d releasenotes/build/doctrees --keep-going -b html releasenotes/source releasenotes/build/html
|
|
||||||
|
|
||||||
[testenv:bandit]
|
|
||||||
deps = -r{toxinidir}/test-requirements.txt
|
|
||||||
commands = bandit -r watcher -x watcher/tests/* -n5 -ll -s B320
|
|
||||||
|
|
||||||
[testenv:lower-constraints]
|
|
||||||
deps =
|
|
||||||
-c{toxinidir}/lower-constraints.txt
|
|
||||||
-r{toxinidir}/test-requirements.txt
|
|
||||||
-r{toxinidir}/requirements.txt
|
|
||||||
|
|||||||
@@ -13,8 +13,6 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
from __future__ import unicode_literals
|
|
||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from watcher.api import hooks
|
from watcher.api import hooks
|
||||||
|
|
||||||
|
|||||||
@@ -57,6 +57,7 @@ are dynamically loaded by Watcher at launch time.
|
|||||||
|
|
||||||
import datetime
|
import datetime
|
||||||
|
|
||||||
|
from http import HTTPStatus
|
||||||
import pecan
|
import pecan
|
||||||
from pecan import rest
|
from pecan import rest
|
||||||
import wsme
|
import wsme
|
||||||
@@ -362,7 +363,7 @@ class ActionsController(rest.RestController):
|
|||||||
|
|
||||||
return Action.convert_with_links(action)
|
return Action.convert_with_links(action)
|
||||||
|
|
||||||
@wsme_pecan.wsexpose(Action, body=Action, status_code=201)
|
@wsme_pecan.wsexpose(Action, body=Action, status_code=HTTPStatus.CREATED)
|
||||||
def post(self, action):
|
def post(self, action):
|
||||||
"""Create a new action(forbidden).
|
"""Create a new action(forbidden).
|
||||||
|
|
||||||
@@ -422,7 +423,7 @@ class ActionsController(rest.RestController):
|
|||||||
action_to_update.save()
|
action_to_update.save()
|
||||||
return Action.convert_with_links(action_to_update)
|
return Action.convert_with_links(action_to_update)
|
||||||
|
|
||||||
@wsme_pecan.wsexpose(None, types.uuid, status_code=204)
|
@wsme_pecan.wsexpose(None, types.uuid, status_code=HTTPStatus.NO_CONTENT)
|
||||||
def delete(self, action_uuid):
|
def delete(self, action_uuid):
|
||||||
"""Delete a action(forbidden).
|
"""Delete a action(forbidden).
|
||||||
|
|
||||||
|
|||||||
@@ -56,6 +56,7 @@ state machine <action_plan_state_machine>`.
|
|||||||
|
|
||||||
import datetime
|
import datetime
|
||||||
|
|
||||||
|
from http import HTTPStatus
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
import pecan
|
import pecan
|
||||||
from pecan import rest
|
from pecan import rest
|
||||||
@@ -460,7 +461,7 @@ class ActionPlansController(rest.RestController):
|
|||||||
|
|
||||||
return ActionPlan.convert_with_links(action_plan)
|
return ActionPlan.convert_with_links(action_plan)
|
||||||
|
|
||||||
@wsme_pecan.wsexpose(None, types.uuid, status_code=204)
|
@wsme_pecan.wsexpose(None, types.uuid, status_code=HTTPStatus.NO_CONTENT)
|
||||||
def delete(self, action_plan_uuid):
|
def delete(self, action_plan_uuid):
|
||||||
"""Delete an action plan.
|
"""Delete an action plan.
|
||||||
|
|
||||||
|
|||||||
@@ -32,6 +32,7 @@ states, visit :ref:`the Audit State machine <audit_state_machine>`.
|
|||||||
import datetime
|
import datetime
|
||||||
from dateutil import tz
|
from dateutil import tz
|
||||||
|
|
||||||
|
from http import HTTPStatus
|
||||||
import pecan
|
import pecan
|
||||||
from pecan import rest
|
from pecan import rest
|
||||||
import wsme
|
import wsme
|
||||||
@@ -595,7 +596,8 @@ class AuditsController(rest.RestController):
|
|||||||
|
|
||||||
return Audit.convert_with_links(rpc_audit)
|
return Audit.convert_with_links(rpc_audit)
|
||||||
|
|
||||||
@wsme_pecan.wsexpose(Audit, body=AuditPostType, status_code=201)
|
@wsme_pecan.wsexpose(Audit, body=AuditPostType,
|
||||||
|
status_code=HTTPStatus.CREATED)
|
||||||
def post(self, audit_p):
|
def post(self, audit_p):
|
||||||
"""Create a new audit.
|
"""Create a new audit.
|
||||||
|
|
||||||
@@ -717,7 +719,7 @@ class AuditsController(rest.RestController):
|
|||||||
audit_to_update.save()
|
audit_to_update.save()
|
||||||
return Audit.convert_with_links(audit_to_update)
|
return Audit.convert_with_links(audit_to_update)
|
||||||
|
|
||||||
@wsme_pecan.wsexpose(None, wtypes.text, status_code=204)
|
@wsme_pecan.wsexpose(None, wtypes.text, status_code=HTTPStatus.NO_CONTENT)
|
||||||
def delete(self, audit):
|
def delete(self, audit):
|
||||||
"""Delete an audit.
|
"""Delete an audit.
|
||||||
|
|
||||||
|
|||||||
@@ -45,6 +45,7 @@ will be launched automatically or will need a manual confirmation from the
|
|||||||
|
|
||||||
import datetime
|
import datetime
|
||||||
|
|
||||||
|
from http import HTTPStatus
|
||||||
import pecan
|
import pecan
|
||||||
from pecan import rest
|
from pecan import rest
|
||||||
import wsme
|
import wsme
|
||||||
@@ -138,6 +139,9 @@ class AuditTemplatePostType(wtypes.Base):
|
|||||||
raise exception.InvalidGoal(goal=audit_template.goal)
|
raise exception.InvalidGoal(goal=audit_template.goal)
|
||||||
|
|
||||||
if audit_template.scope:
|
if audit_template.scope:
|
||||||
|
keys = [list(s)[0] for s in audit_template.scope]
|
||||||
|
if keys[0] not in ('compute', 'storage'):
|
||||||
|
audit_template.scope = [dict(compute=audit_template.scope)]
|
||||||
common_utils.Draft4Validator(
|
common_utils.Draft4Validator(
|
||||||
AuditTemplatePostType._build_schema()
|
AuditTemplatePostType._build_schema()
|
||||||
).validate(audit_template.scope)
|
).validate(audit_template.scope)
|
||||||
@@ -158,18 +162,23 @@ class AuditTemplatePostType(wtypes.Base):
|
|||||||
"included and excluded together"))
|
"included and excluded together"))
|
||||||
|
|
||||||
if audit_template.strategy:
|
if audit_template.strategy:
|
||||||
available_strategies = objects.Strategy.list(
|
try:
|
||||||
AuditTemplatePostType._ctx)
|
if (common_utils.is_uuid_like(audit_template.strategy) or
|
||||||
available_strategies_map = {
|
common_utils.is_int_like(audit_template.strategy)):
|
||||||
s.uuid: s for s in available_strategies}
|
strategy = objects.Strategy.get(
|
||||||
if audit_template.strategy not in available_strategies_map:
|
AuditTemplatePostType._ctx, audit_template.strategy)
|
||||||
|
else:
|
||||||
|
strategy = objects.Strategy.get_by_name(
|
||||||
|
AuditTemplatePostType._ctx, audit_template.strategy)
|
||||||
|
except Exception:
|
||||||
raise exception.InvalidStrategy(
|
raise exception.InvalidStrategy(
|
||||||
strategy=audit_template.strategy)
|
strategy=audit_template.strategy)
|
||||||
|
|
||||||
strategy = available_strategies_map[audit_template.strategy]
|
|
||||||
# Check that the strategy we indicate is actually related to the
|
# Check that the strategy we indicate is actually related to the
|
||||||
# specified goal
|
# specified goal
|
||||||
if strategy.goal_id != goal.id:
|
if strategy.goal_id != goal.id:
|
||||||
|
available_strategies = objects.Strategy.list(
|
||||||
|
AuditTemplatePostType._ctx)
|
||||||
choices = ["'%s' (%s)" % (s.uuid, s.name)
|
choices = ["'%s' (%s)" % (s.uuid, s.name)
|
||||||
for s in available_strategies]
|
for s in available_strategies]
|
||||||
raise exception.InvalidStrategy(
|
raise exception.InvalidStrategy(
|
||||||
@@ -610,7 +619,7 @@ class AuditTemplatesController(rest.RestController):
|
|||||||
|
|
||||||
@wsme.validate(types.uuid, AuditTemplatePostType)
|
@wsme.validate(types.uuid, AuditTemplatePostType)
|
||||||
@wsme_pecan.wsexpose(AuditTemplate, body=AuditTemplatePostType,
|
@wsme_pecan.wsexpose(AuditTemplate, body=AuditTemplatePostType,
|
||||||
status_code=201)
|
status_code=HTTPStatus.CREATED)
|
||||||
def post(self, audit_template_postdata):
|
def post(self, audit_template_postdata):
|
||||||
"""Create a new audit template.
|
"""Create a new audit template.
|
||||||
|
|
||||||
@@ -686,7 +695,7 @@ class AuditTemplatesController(rest.RestController):
|
|||||||
audit_template_to_update.save()
|
audit_template_to_update.save()
|
||||||
return AuditTemplate.convert_with_links(audit_template_to_update)
|
return AuditTemplate.convert_with_links(audit_template_to_update)
|
||||||
|
|
||||||
@wsme_pecan.wsexpose(None, wtypes.text, status_code=204)
|
@wsme_pecan.wsexpose(None, wtypes.text, status_code=HTTPStatus.NO_CONTENT)
|
||||||
def delete(self, audit_template):
|
def delete(self, audit_template):
|
||||||
"""Delete a audit template.
|
"""Delete a audit template.
|
||||||
|
|
||||||
|
|||||||
@@ -19,8 +19,6 @@ Service mechanism provides ability to monitor Watcher services state.
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
import datetime
|
import datetime
|
||||||
import six
|
|
||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
@@ -70,7 +68,7 @@ class Service(base.APIBase):
|
|||||||
service = objects.Service.get(pecan.request.context, id)
|
service = objects.Service.get(pecan.request.context, id)
|
||||||
last_heartbeat = (service.last_seen_up or service.updated_at or
|
last_heartbeat = (service.last_seen_up or service.updated_at or
|
||||||
service.created_at)
|
service.created_at)
|
||||||
if isinstance(last_heartbeat, six.string_types):
|
if isinstance(last_heartbeat, str):
|
||||||
# NOTE(russellb) If this service came in over rpc via
|
# NOTE(russellb) If this service came in over rpc via
|
||||||
# conductor, then the timestamp will be a string and needs to be
|
# conductor, then the timestamp will be a string and needs to be
|
||||||
# converted back to a datetime.
|
# converted back to a datetime.
|
||||||
|
|||||||
@@ -15,7 +15,6 @@
|
|||||||
|
|
||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
from oslo_utils import strutils
|
from oslo_utils import strutils
|
||||||
import six
|
|
||||||
import wsme
|
import wsme
|
||||||
from wsme import types as wtypes
|
from wsme import types as wtypes
|
||||||
|
|
||||||
@@ -132,7 +131,7 @@ class JsonType(wtypes.UserType):
|
|||||||
|
|
||||||
def __str__(self):
|
def __str__(self):
|
||||||
# These are the json serializable native types
|
# These are the json serializable native types
|
||||||
return ' | '.join(map(str, (wtypes.text, six.integer_types, float,
|
return ' | '.join(map(str, (wtypes.text, int, float,
|
||||||
BooleanType, list, dict, None)))
|
BooleanType, list, dict, None)))
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
|||||||
@@ -14,6 +14,7 @@
|
|||||||
Webhook endpoint for Watcher v1 REST API.
|
Webhook endpoint for Watcher v1 REST API.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
from http import HTTPStatus
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
import pecan
|
import pecan
|
||||||
from pecan import rest
|
from pecan import rest
|
||||||
@@ -36,7 +37,7 @@ class WebhookController(rest.RestController):
|
|||||||
self.dc_client = rpcapi.DecisionEngineAPI()
|
self.dc_client = rpcapi.DecisionEngineAPI()
|
||||||
|
|
||||||
@wsme_pecan.wsexpose(None, wtypes.text, body=types.jsontype,
|
@wsme_pecan.wsexpose(None, wtypes.text, body=types.jsontype,
|
||||||
status_code=202)
|
status_code=HTTPStatus.ACCEPTED)
|
||||||
def post(self, audit_ident, body):
|
def post(self, audit_ident, body):
|
||||||
"""Trigger the given audit.
|
"""Trigger the given audit.
|
||||||
|
|
||||||
|
|||||||
@@ -15,9 +15,9 @@
|
|||||||
# under the License.
|
# under the License.
|
||||||
|
|
||||||
|
|
||||||
|
from http import HTTPStatus
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from pecan import hooks
|
from pecan import hooks
|
||||||
from six.moves import http_client
|
|
||||||
|
|
||||||
from watcher.common import context
|
from watcher.common import context
|
||||||
|
|
||||||
@@ -91,8 +91,8 @@ class NoExceptionTracebackHook(hooks.PecanHook):
|
|||||||
# Do nothing if there is no error.
|
# Do nothing if there is no error.
|
||||||
# Status codes in the range 200 (OK) to 399 (400 = BAD_REQUEST) are not
|
# Status codes in the range 200 (OK) to 399 (400 = BAD_REQUEST) are not
|
||||||
# an error.
|
# an error.
|
||||||
if (http_client.OK <= state.response.status_int <
|
if (HTTPStatus.OK <= state.response.status_int <
|
||||||
http_client.BAD_REQUEST):
|
HTTPStatus.BAD_REQUEST):
|
||||||
return
|
return
|
||||||
|
|
||||||
json_body = state.response.json
|
json_body = state.response.json
|
||||||
|
|||||||
@@ -24,7 +24,6 @@ from xml import etree as et
|
|||||||
|
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
import six
|
|
||||||
import webob
|
import webob
|
||||||
|
|
||||||
from watcher._i18n import _
|
from watcher._i18n import _
|
||||||
@@ -84,12 +83,10 @@ class ParsableErrorMiddleware(object):
|
|||||||
'</error_message>' % state['status_code']]
|
'</error_message>' % state['status_code']]
|
||||||
state['headers'].append(('Content-Type', 'application/xml'))
|
state['headers'].append(('Content-Type', 'application/xml'))
|
||||||
else:
|
else:
|
||||||
if six.PY3:
|
app_iter = [i.decode('utf-8') for i in app_iter]
|
||||||
app_iter = [i.decode('utf-8') for i in app_iter]
|
|
||||||
body = [jsonutils.dumps(
|
body = [jsonutils.dumps(
|
||||||
{'error_message': '\n'.join(app_iter)})]
|
{'error_message': '\n'.join(app_iter)})]
|
||||||
if six.PY3:
|
body = [item.encode('utf-8') for item in body]
|
||||||
body = [item.encode('utf-8') for item in body]
|
|
||||||
state['headers'].append(('Content-Type', 'application/json'))
|
state['headers'].append(('Content-Type', 'application/json'))
|
||||||
state['headers'].append(('Content-Length', str(len(body[0]))))
|
state['headers'].append(('Content-Length', str(len(body[0]))))
|
||||||
else:
|
else:
|
||||||
|
|||||||
@@ -20,7 +20,6 @@ import itertools
|
|||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
import six
|
|
||||||
|
|
||||||
from watcher.common import context as watcher_context
|
from watcher.common import context as watcher_context
|
||||||
from watcher.common import scheduling
|
from watcher.common import scheduling
|
||||||
@@ -83,7 +82,7 @@ class APISchedulingService(scheduling.BackgroundSchedulerService):
|
|||||||
service = objects.Service.get(context, service_id)
|
service = objects.Service.get(context, service_id)
|
||||||
last_heartbeat = (service.last_seen_up or service.updated_at or
|
last_heartbeat = (service.last_seen_up or service.updated_at or
|
||||||
service.created_at)
|
service.created_at)
|
||||||
if isinstance(last_heartbeat, six.string_types):
|
if isinstance(last_heartbeat, str):
|
||||||
# NOTE(russellb) If this service came in over rpc via
|
# NOTE(russellb) If this service came in over rpc via
|
||||||
# conductor, then the timestamp will be a string and needs to be
|
# conductor, then the timestamp will be a string and needs to be
|
||||||
# converted back to a datetime.
|
# converted back to a datetime.
|
||||||
|
|||||||
@@ -18,11 +18,9 @@
|
|||||||
#
|
#
|
||||||
|
|
||||||
import abc
|
import abc
|
||||||
import six
|
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class BaseActionPlanHandler(object, metaclass=abc.ABCMeta):
|
||||||
class BaseActionPlanHandler(object):
|
|
||||||
@abc.abstractmethod
|
@abc.abstractmethod
|
||||||
def execute(self):
|
def execute(self):
|
||||||
raise NotImplementedError()
|
raise NotImplementedError()
|
||||||
|
|||||||
@@ -19,14 +19,12 @@
|
|||||||
import abc
|
import abc
|
||||||
|
|
||||||
import jsonschema
|
import jsonschema
|
||||||
import six
|
|
||||||
|
|
||||||
from watcher.common import clients
|
from watcher.common import clients
|
||||||
from watcher.common.loader import loadable
|
from watcher.common.loader import loadable
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class BaseAction(loadable.Loadable, metaclass=abc.ABCMeta):
|
||||||
class BaseAction(loadable.Loadable):
|
|
||||||
# NOTE(jed): by convention we decided
|
# NOTE(jed): by convention we decided
|
||||||
# that the attribute "resource_id" is the unique id of
|
# that the attribute "resource_id" is the unique id of
|
||||||
# the resource to which the Action applies to allow us to use it in the
|
# the resource to which the Action applies to allow us to use it in the
|
||||||
|
|||||||
@@ -17,17 +17,17 @@
|
|||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
#
|
#
|
||||||
|
|
||||||
import enum
|
|
||||||
import time
|
import time
|
||||||
|
|
||||||
|
from oslo_log import log
|
||||||
|
|
||||||
from watcher._i18n import _
|
from watcher._i18n import _
|
||||||
from watcher.applier.actions import base
|
from watcher.applier.actions import base
|
||||||
from watcher.common import exception
|
from watcher.common import exception
|
||||||
|
from watcher.common.metal_helper import constants as metal_constants
|
||||||
|
from watcher.common.metal_helper import factory as metal_helper_factory
|
||||||
|
|
||||||
|
LOG = log.getLogger(__name__)
|
||||||
class NodeState(enum.Enum):
|
|
||||||
POWERON = 'on'
|
|
||||||
POWEROFF = 'off'
|
|
||||||
|
|
||||||
|
|
||||||
class ChangeNodePowerState(base.BaseAction):
|
class ChangeNodePowerState(base.BaseAction):
|
||||||
@@ -43,8 +43,8 @@ class ChangeNodePowerState(base.BaseAction):
|
|||||||
'state': str,
|
'state': str,
|
||||||
})
|
})
|
||||||
|
|
||||||
The `resource_id` references a ironic node id (list of available
|
The `resource_id` references a baremetal node id (list of available
|
||||||
ironic node is returned by this command: ``ironic node-list``).
|
ironic nodes is returned by this command: ``ironic node-list``).
|
||||||
The `state` value should either be `on` or `off`.
|
The `state` value should either be `on` or `off`.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
@@ -59,10 +59,14 @@ class ChangeNodePowerState(base.BaseAction):
|
|||||||
'type': 'string',
|
'type': 'string',
|
||||||
"minlength": 1
|
"minlength": 1
|
||||||
},
|
},
|
||||||
|
'resource_name': {
|
||||||
|
'type': 'string',
|
||||||
|
"minlength": 1
|
||||||
|
},
|
||||||
'state': {
|
'state': {
|
||||||
'type': 'string',
|
'type': 'string',
|
||||||
'enum': [NodeState.POWERON.value,
|
'enum': [metal_constants.PowerState.ON.value,
|
||||||
NodeState.POWEROFF.value]
|
metal_constants.PowerState.OFF.value]
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
'required': ['resource_id', 'state'],
|
'required': ['resource_id', 'state'],
|
||||||
@@ -82,10 +86,10 @@ class ChangeNodePowerState(base.BaseAction):
|
|||||||
return self._node_manage_power(target_state)
|
return self._node_manage_power(target_state)
|
||||||
|
|
||||||
def revert(self):
|
def revert(self):
|
||||||
if self.state == NodeState.POWERON.value:
|
if self.state == metal_constants.PowerState.ON.value:
|
||||||
target_state = NodeState.POWEROFF.value
|
target_state = metal_constants.PowerState.OFF.value
|
||||||
elif self.state == NodeState.POWEROFF.value:
|
elif self.state == metal_constants.PowerState.OFF.value:
|
||||||
target_state = NodeState.POWERON.value
|
target_state = metal_constants.PowerState.ON.value
|
||||||
return self._node_manage_power(target_state)
|
return self._node_manage_power(target_state)
|
||||||
|
|
||||||
def _node_manage_power(self, state, retry=60):
|
def _node_manage_power(self, state, retry=60):
|
||||||
@@ -93,30 +97,32 @@ class ChangeNodePowerState(base.BaseAction):
|
|||||||
raise exception.IllegalArgumentException(
|
raise exception.IllegalArgumentException(
|
||||||
message=_("The target state is not defined"))
|
message=_("The target state is not defined"))
|
||||||
|
|
||||||
ironic_client = self.osc.ironic()
|
metal_helper = metal_helper_factory.get_helper(self.osc)
|
||||||
nova_client = self.osc.nova()
|
node = metal_helper.get_node(self.node_uuid)
|
||||||
current_state = ironic_client.node.get(self.node_uuid).power_state
|
current_state = node.get_power_state()
|
||||||
# power state: 'power on' or 'power off', if current node state
|
|
||||||
# is the same as state, just return True
|
if state == current_state.value:
|
||||||
if state in current_state:
|
|
||||||
return True
|
return True
|
||||||
|
|
||||||
if state == NodeState.POWEROFF.value:
|
if state == metal_constants.PowerState.OFF.value:
|
||||||
node_info = ironic_client.node.get(self.node_uuid).to_dict()
|
compute_node = node.get_hypervisor_node().to_dict()
|
||||||
compute_node_id = node_info['extra']['compute_node_id']
|
|
||||||
compute_node = nova_client.hypervisors.get(compute_node_id)
|
|
||||||
compute_node = compute_node.to_dict()
|
|
||||||
if (compute_node['running_vms'] == 0):
|
if (compute_node['running_vms'] == 0):
|
||||||
ironic_client.node.set_power_state(
|
node.set_power_state(state)
|
||||||
self.node_uuid, state)
|
else:
|
||||||
|
LOG.warning(
|
||||||
|
"Compute node %s has %s running vms and will "
|
||||||
|
"NOT be shut off.",
|
||||||
|
compute_node["hypervisor_hostname"],
|
||||||
|
compute_node['running_vms'])
|
||||||
|
return False
|
||||||
else:
|
else:
|
||||||
ironic_client.node.set_power_state(self.node_uuid, state)
|
node.set_power_state(state)
|
||||||
|
|
||||||
ironic_node = ironic_client.node.get(self.node_uuid)
|
node = metal_helper.get_node(self.node_uuid)
|
||||||
while ironic_node.power_state == current_state and retry:
|
while node.get_power_state() == current_state and retry:
|
||||||
time.sleep(10)
|
time.sleep(10)
|
||||||
retry -= 1
|
retry -= 1
|
||||||
ironic_node = ironic_client.node.get(self.node_uuid)
|
node = metal_helper.get_node(self.node_uuid)
|
||||||
if retry > 0:
|
if retry > 0:
|
||||||
return True
|
return True
|
||||||
else:
|
else:
|
||||||
@@ -130,4 +136,4 @@ class ChangeNodePowerState(base.BaseAction):
|
|||||||
|
|
||||||
def get_description(self):
|
def get_description(self):
|
||||||
"""Description of the action"""
|
"""Description of the action"""
|
||||||
return ("Compute node power on/off through ironic.")
|
return ("Compute node power on/off through Ironic or MaaS.")
|
||||||
|
|||||||
@@ -15,8 +15,6 @@
|
|||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
#
|
#
|
||||||
|
|
||||||
from __future__ import unicode_literals
|
|
||||||
|
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
|
|
||||||
from watcher.applier.loading import default
|
from watcher.applier.loading import default
|
||||||
|
|||||||
@@ -186,7 +186,7 @@ class Migrate(base.BaseAction):
|
|||||||
return self.migrate(destination=self.destination_node)
|
return self.migrate(destination=self.destination_node)
|
||||||
|
|
||||||
def revert(self):
|
def revert(self):
|
||||||
LOG.info('Migrate action do not revert!')
|
return self.migrate(destination=self.source_node)
|
||||||
|
|
||||||
def abort(self):
|
def abort(self):
|
||||||
nova = nova_helper.NovaHelper(osc=self.osc)
|
nova = nova_helper.NovaHelper(osc=self.osc)
|
||||||
|
|||||||
@@ -95,7 +95,7 @@ class Resize(base.BaseAction):
|
|||||||
return self.resize()
|
return self.resize()
|
||||||
|
|
||||||
def revert(self):
|
def revert(self):
|
||||||
return self.migrate(destination=self.source_node)
|
LOG.warning("revert not supported")
|
||||||
|
|
||||||
def pre_condition(self):
|
def pre_condition(self):
|
||||||
# TODO(jed): check if the instance exists / check if the instance is on
|
# TODO(jed): check if the instance exists / check if the instance is on
|
||||||
|
|||||||
@@ -26,11 +26,9 @@ See: :doc:`../architecture` for more details on this component.
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
import abc
|
import abc
|
||||||
import six
|
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class BaseApplier(object, metaclass=abc.ABCMeta):
|
||||||
class BaseApplier(object):
|
|
||||||
@abc.abstractmethod
|
@abc.abstractmethod
|
||||||
def execute(self, action_plan_uuid):
|
def execute(self, action_plan_uuid):
|
||||||
raise NotImplementedError()
|
raise NotImplementedError()
|
||||||
|
|||||||
@@ -11,9 +11,6 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
from __future__ import unicode_literals
|
|
||||||
|
|
||||||
|
|
||||||
from watcher.common.loader import default
|
from watcher.common.loader import default
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -17,7 +17,6 @@
|
|||||||
#
|
#
|
||||||
|
|
||||||
import abc
|
import abc
|
||||||
import six
|
|
||||||
import time
|
import time
|
||||||
|
|
||||||
import eventlet
|
import eventlet
|
||||||
@@ -40,8 +39,7 @@ CANCEL_STATE = [objects.action_plan.State.CANCELLING,
|
|||||||
objects.action_plan.State.CANCELLED]
|
objects.action_plan.State.CANCELLED]
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class BaseWorkFlowEngine(loadable.Loadable, metaclass=abc.ABCMeta):
|
||||||
class BaseWorkFlowEngine(loadable.Loadable):
|
|
||||||
|
|
||||||
def __init__(self, config, context=None, applier_manager=None):
|
def __init__(self, config, context=None, applier_manager=None):
|
||||||
"""Constructor
|
"""Constructor
|
||||||
|
|||||||
@@ -25,8 +25,11 @@ from taskflow import task as flow_task
|
|||||||
|
|
||||||
from watcher.applier.workflow_engine import base
|
from watcher.applier.workflow_engine import base
|
||||||
from watcher.common import exception
|
from watcher.common import exception
|
||||||
|
from watcher import conf
|
||||||
from watcher import objects
|
from watcher import objects
|
||||||
|
|
||||||
|
CONF = conf.CONF
|
||||||
|
|
||||||
LOG = log.getLogger(__name__)
|
LOG = log.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
@@ -127,9 +130,11 @@ class DefaultWorkFlowEngine(base.BaseWorkFlowEngine):
|
|||||||
|
|
||||||
class TaskFlowActionContainer(base.BaseTaskFlowActionContainer):
|
class TaskFlowActionContainer(base.BaseTaskFlowActionContainer):
|
||||||
def __init__(self, db_action, engine):
|
def __init__(self, db_action, engine):
|
||||||
name = "action_type:{0} uuid:{1}".format(db_action.action_type,
|
self.name = "action_type:{0} uuid:{1}".format(db_action.action_type,
|
||||||
db_action.uuid)
|
db_action.uuid)
|
||||||
super(TaskFlowActionContainer, self).__init__(name, db_action, engine)
|
super(TaskFlowActionContainer, self).__init__(self.name,
|
||||||
|
db_action,
|
||||||
|
engine)
|
||||||
|
|
||||||
def do_pre_execute(self):
|
def do_pre_execute(self):
|
||||||
db_action = self.engine.notify(self._db_action,
|
db_action = self.engine.notify(self._db_action,
|
||||||
@@ -158,6 +163,12 @@ class TaskFlowActionContainer(base.BaseTaskFlowActionContainer):
|
|||||||
self.action.post_condition()
|
self.action.post_condition()
|
||||||
|
|
||||||
def do_revert(self, *args, **kwargs):
|
def do_revert(self, *args, **kwargs):
|
||||||
|
# NOTE: Not rollback action plan
|
||||||
|
if not CONF.watcher_applier.rollback_when_actionplan_failed:
|
||||||
|
LOG.info("Failed actionplan rollback option is turned off, and "
|
||||||
|
"the following action will be skipped: %s", self.name)
|
||||||
|
return
|
||||||
|
|
||||||
LOG.warning("Revert action: %s", self.name)
|
LOG.warning("Revert action: %s", self.name)
|
||||||
try:
|
try:
|
||||||
# TODO(jed): do we need to update the states in case of failure?
|
# TODO(jed): do we need to update the states in case of failure?
|
||||||
|
|||||||
@@ -18,3 +18,10 @@
|
|||||||
import eventlet
|
import eventlet
|
||||||
|
|
||||||
eventlet.monkey_patch()
|
eventlet.monkey_patch()
|
||||||
|
|
||||||
|
# Monkey patch the original current_thread to use the up-to-date _active
|
||||||
|
# global variable. See https://bugs.launchpad.net/bugs/1863021 and
|
||||||
|
# https://github.com/eventlet/eventlet/issues/592
|
||||||
|
import __original_module_threading as orig_threading # noqa
|
||||||
|
import threading # noqa
|
||||||
|
orig_threading.current_thread.__globals__['_active'] = threading._active
|
||||||
|
|||||||
@@ -14,6 +14,7 @@
|
|||||||
|
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
|
from oslo_upgradecheck import common_checks
|
||||||
from oslo_upgradecheck import upgradecheck
|
from oslo_upgradecheck import upgradecheck
|
||||||
|
|
||||||
from watcher._i18n import _
|
from watcher._i18n import _
|
||||||
@@ -43,6 +44,10 @@ class Checks(upgradecheck.UpgradeCommands):
|
|||||||
_upgrade_checks = (
|
_upgrade_checks = (
|
||||||
# Added in Train.
|
# Added in Train.
|
||||||
(_('Minimum Nova API Version'), _minimum_nova_api_version),
|
(_('Minimum Nova API Version'), _minimum_nova_api_version),
|
||||||
|
# Added in Wallaby.
|
||||||
|
(_("Policy File JSON to YAML Migration"),
|
||||||
|
(common_checks.check_policy_json, {'conf': CONF})),
|
||||||
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -17,7 +17,7 @@ import time
|
|||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
|
|
||||||
from cinderclient import exceptions as cinder_exception
|
from cinderclient import exceptions as cinder_exception
|
||||||
from cinderclient.v2.volumes import Volume
|
from cinderclient.v3.volumes import Volume
|
||||||
from watcher._i18n import _
|
from watcher._i18n import _
|
||||||
from watcher.common import clients
|
from watcher.common import clients
|
||||||
from watcher.common import exception
|
from watcher.common import exception
|
||||||
|
|||||||
@@ -25,6 +25,7 @@ from novaclient import api_versions as nova_api_versions
|
|||||||
from novaclient import client as nvclient
|
from novaclient import client as nvclient
|
||||||
|
|
||||||
from watcher.common import exception
|
from watcher.common import exception
|
||||||
|
from watcher.common import utils
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from ceilometerclient import client as ceclient
|
from ceilometerclient import client as ceclient
|
||||||
@@ -32,6 +33,12 @@ try:
|
|||||||
except ImportError:
|
except ImportError:
|
||||||
HAS_CEILCLIENT = False
|
HAS_CEILCLIENT = False
|
||||||
|
|
||||||
|
try:
|
||||||
|
from maas import client as maas_client
|
||||||
|
except ImportError:
|
||||||
|
maas_client = None
|
||||||
|
|
||||||
|
|
||||||
CONF = cfg.CONF
|
CONF = cfg.CONF
|
||||||
|
|
||||||
_CLIENTS_AUTH_GROUP = 'watcher_clients_auth'
|
_CLIENTS_AUTH_GROUP = 'watcher_clients_auth'
|
||||||
@@ -74,6 +81,7 @@ class OpenStackClients(object):
|
|||||||
self._monasca = None
|
self._monasca = None
|
||||||
self._neutron = None
|
self._neutron = None
|
||||||
self._ironic = None
|
self._ironic = None
|
||||||
|
self._maas = None
|
||||||
self._placement = None
|
self._placement = None
|
||||||
|
|
||||||
def _get_keystone_session(self):
|
def _get_keystone_session(self):
|
||||||
@@ -265,6 +273,23 @@ class OpenStackClients(object):
|
|||||||
session=self.session)
|
session=self.session)
|
||||||
return self._ironic
|
return self._ironic
|
||||||
|
|
||||||
|
def maas(self):
|
||||||
|
if self._maas:
|
||||||
|
return self._maas
|
||||||
|
|
||||||
|
if not maas_client:
|
||||||
|
raise exception.UnsupportedError(
|
||||||
|
"MAAS client unavailable. Please install python-libmaas.")
|
||||||
|
|
||||||
|
url = self._get_client_option('maas', 'url')
|
||||||
|
api_key = self._get_client_option('maas', 'api_key')
|
||||||
|
timeout = self._get_client_option('maas', 'timeout')
|
||||||
|
self._maas = utils.async_compat_call(
|
||||||
|
maas_client.connect,
|
||||||
|
url, apikey=api_key,
|
||||||
|
timeout=timeout)
|
||||||
|
return self._maas
|
||||||
|
|
||||||
@exception.wrap_keystone_exception
|
@exception.wrap_keystone_exception
|
||||||
def placement(self):
|
def placement(self):
|
||||||
if self._placement:
|
if self._placement:
|
||||||
|
|||||||
@@ -13,7 +13,6 @@
|
|||||||
from oslo_context import context
|
from oslo_context import context
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
import six
|
|
||||||
|
|
||||||
LOG = log.getLogger(__name__)
|
LOG = log.getLogger(__name__)
|
||||||
|
|
||||||
@@ -69,7 +68,7 @@ class RequestContext(context.RequestContext):
|
|||||||
self.project_id = project_id
|
self.project_id = project_id
|
||||||
if not timestamp:
|
if not timestamp:
|
||||||
timestamp = timeutils.utcnow()
|
timestamp = timeutils.utcnow()
|
||||||
if isinstance(timestamp, six.string_types):
|
if isinstance(timestamp, str):
|
||||||
timestamp = timeutils.parse_isotime(timestamp)
|
timestamp = timeutils.parse_isotime(timestamp)
|
||||||
self.timestamp = timestamp
|
self.timestamp = timestamp
|
||||||
self.user_name = user_name
|
self.user_name = user_name
|
||||||
|
|||||||
@@ -25,6 +25,7 @@ SHOULD include dedicated exception logging.
|
|||||||
import functools
|
import functools
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
|
from http import HTTPStatus
|
||||||
from keystoneclient import exceptions as keystone_exceptions
|
from keystoneclient import exceptions as keystone_exceptions
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
@@ -62,7 +63,7 @@ class WatcherException(Exception):
|
|||||||
|
|
||||||
"""
|
"""
|
||||||
msg_fmt = _("An unknown exception occurred")
|
msg_fmt = _("An unknown exception occurred")
|
||||||
code = 500
|
code = HTTPStatus.INTERNAL_SERVER_ERROR
|
||||||
headers = {}
|
headers = {}
|
||||||
safe = False
|
safe = False
|
||||||
|
|
||||||
@@ -114,12 +115,12 @@ class UnsupportedError(WatcherException):
|
|||||||
|
|
||||||
class NotAuthorized(WatcherException):
|
class NotAuthorized(WatcherException):
|
||||||
msg_fmt = _("Not authorized")
|
msg_fmt = _("Not authorized")
|
||||||
code = 403
|
code = HTTPStatus.FORBIDDEN
|
||||||
|
|
||||||
|
|
||||||
class NotAcceptable(WatcherException):
|
class NotAcceptable(WatcherException):
|
||||||
msg_fmt = _("Request not acceptable.")
|
msg_fmt = _("Request not acceptable.")
|
||||||
code = 406
|
code = HTTPStatus.NOT_ACCEPTABLE
|
||||||
|
|
||||||
|
|
||||||
class PolicyNotAuthorized(NotAuthorized):
|
class PolicyNotAuthorized(NotAuthorized):
|
||||||
@@ -132,7 +133,7 @@ class OperationNotPermitted(NotAuthorized):
|
|||||||
|
|
||||||
class Invalid(WatcherException, ValueError):
|
class Invalid(WatcherException, ValueError):
|
||||||
msg_fmt = _("Unacceptable parameters")
|
msg_fmt = _("Unacceptable parameters")
|
||||||
code = 400
|
code = HTTPStatus.BAD_REQUEST
|
||||||
|
|
||||||
|
|
||||||
class ObjectNotFound(WatcherException):
|
class ObjectNotFound(WatcherException):
|
||||||
@@ -141,12 +142,12 @@ class ObjectNotFound(WatcherException):
|
|||||||
|
|
||||||
class Conflict(WatcherException):
|
class Conflict(WatcherException):
|
||||||
msg_fmt = _('Conflict')
|
msg_fmt = _('Conflict')
|
||||||
code = 409
|
code = HTTPStatus.CONFLICT
|
||||||
|
|
||||||
|
|
||||||
class ResourceNotFound(ObjectNotFound):
|
class ResourceNotFound(ObjectNotFound):
|
||||||
msg_fmt = _("The %(name)s resource %(id)s could not be found")
|
msg_fmt = _("The %(name)s resource %(id)s could not be found")
|
||||||
code = 404
|
code = HTTPStatus.NOT_FOUND
|
||||||
|
|
||||||
|
|
||||||
class InvalidParameter(Invalid):
|
class InvalidParameter(Invalid):
|
||||||
|
|||||||
@@ -14,14 +14,10 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
from __future__ import unicode_literals
|
|
||||||
|
|
||||||
import abc
|
import abc
|
||||||
import six
|
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class BaseLoader(object, metaclass=abc.ABCMeta):
|
||||||
class BaseLoader(object):
|
|
||||||
|
|
||||||
@abc.abstractmethod
|
@abc.abstractmethod
|
||||||
def list_available(self):
|
def list_available(self):
|
||||||
|
|||||||
@@ -14,8 +14,6 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
from __future__ import unicode_literals
|
|
||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
from stevedore import driver as drivermanager
|
from stevedore import driver as drivermanager
|
||||||
|
|||||||
@@ -16,13 +16,10 @@
|
|||||||
|
|
||||||
import abc
|
import abc
|
||||||
|
|
||||||
import six
|
|
||||||
|
|
||||||
from watcher.common import service
|
from watcher.common import service
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class Loadable(object, metaclass=abc.ABCMeta):
|
||||||
class Loadable(object):
|
|
||||||
"""Generic interface for dynamically loading a driver/entry point.
|
"""Generic interface for dynamically loading a driver/entry point.
|
||||||
|
|
||||||
This defines the contract in order to let the loader manager inject
|
This defines the contract in order to let the loader manager inject
|
||||||
@@ -48,8 +45,7 @@ LoadableSingletonMeta = type(
|
|||||||
"LoadableSingletonMeta", (abc.ABCMeta, service.Singleton), {})
|
"LoadableSingletonMeta", (abc.ABCMeta, service.Singleton), {})
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(LoadableSingletonMeta)
|
class LoadableSingleton(object, metaclass=LoadableSingletonMeta):
|
||||||
class LoadableSingleton(object):
|
|
||||||
"""Generic interface for dynamically loading a driver as a singleton.
|
"""Generic interface for dynamically loading a driver as a singleton.
|
||||||
|
|
||||||
This defines the contract in order to let the loader manager inject
|
This defines the contract in order to let the loader manager inject
|
||||||
|
|||||||
0
watcher/common/metal_helper/__init__.py
Normal file
0
watcher/common/metal_helper/__init__.py
Normal file
81
watcher/common/metal_helper/base.py
Normal file
81
watcher/common/metal_helper/base.py
Normal file
@@ -0,0 +1,81 @@
|
|||||||
|
# Copyright 2023 Cloudbase Solutions
|
||||||
|
# All Rights Reserved.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||||
|
# not use this file except in compliance with the License. You may obtain
|
||||||
|
# a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||||
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||||
|
# License for the specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
|
||||||
|
import abc
|
||||||
|
|
||||||
|
from watcher.common import exception
|
||||||
|
from watcher.common.metal_helper import constants as metal_constants
|
||||||
|
|
||||||
|
|
||||||
|
class BaseMetalNode(abc.ABC):
|
||||||
|
hv_up_when_powered_off = False
|
||||||
|
|
||||||
|
def __init__(self, nova_node=None):
|
||||||
|
self._nova_node = nova_node
|
||||||
|
|
||||||
|
def get_hypervisor_node(self):
|
||||||
|
if not self._nova_node:
|
||||||
|
raise exception.Invalid(message="No associated hypervisor.")
|
||||||
|
return self._nova_node
|
||||||
|
|
||||||
|
def get_hypervisor_hostname(self):
|
||||||
|
return self.get_hypervisor_node().hypervisor_hostname
|
||||||
|
|
||||||
|
@abc.abstractmethod
|
||||||
|
def get_power_state(self):
|
||||||
|
# TODO(lpetrut): document the following methods
|
||||||
|
pass
|
||||||
|
|
||||||
|
@abc.abstractmethod
|
||||||
|
def get_id(self):
|
||||||
|
"""Return the node id provided by the bare metal service."""
|
||||||
|
pass
|
||||||
|
|
||||||
|
@abc.abstractmethod
|
||||||
|
def power_on(self):
|
||||||
|
pass
|
||||||
|
|
||||||
|
@abc.abstractmethod
|
||||||
|
def power_off(self):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def set_power_state(self, state):
|
||||||
|
state = metal_constants.PowerState(state)
|
||||||
|
if state == metal_constants.PowerState.ON:
|
||||||
|
self.power_on()
|
||||||
|
elif state == metal_constants.PowerState.OFF:
|
||||||
|
self.power_off()
|
||||||
|
else:
|
||||||
|
raise exception.UnsupportedActionType(
|
||||||
|
"Cannot set power state: %s" % state)
|
||||||
|
|
||||||
|
|
||||||
|
class BaseMetalHelper(abc.ABC):
|
||||||
|
def __init__(self, osc):
|
||||||
|
self._osc = osc
|
||||||
|
|
||||||
|
@property
|
||||||
|
def nova_client(self):
|
||||||
|
if not getattr(self, "_nova_client", None):
|
||||||
|
self._nova_client = self._osc.nova()
|
||||||
|
return self._nova_client
|
||||||
|
|
||||||
|
@abc.abstractmethod
|
||||||
|
def list_compute_nodes(self):
|
||||||
|
pass
|
||||||
|
|
||||||
|
@abc.abstractmethod
|
||||||
|
def get_node(self, node_id):
|
||||||
|
pass
|
||||||
23
watcher/common/metal_helper/constants.py
Normal file
23
watcher/common/metal_helper/constants.py
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
# Copyright 2023 Cloudbase Solutions
|
||||||
|
# All Rights Reserved.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||||
|
# not use this file except in compliance with the License. You may obtain
|
||||||
|
# a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||||
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||||
|
# License for the specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
|
||||||
|
import enum
|
||||||
|
|
||||||
|
|
||||||
|
class PowerState(str, enum.Enum):
|
||||||
|
ON = "on"
|
||||||
|
OFF = "off"
|
||||||
|
UNKNOWN = "unknown"
|
||||||
|
ERROR = "error"
|
||||||
33
watcher/common/metal_helper/factory.py
Normal file
33
watcher/common/metal_helper/factory.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
# Copyright 2023 Cloudbase Solutions
|
||||||
|
# All Rights Reserved.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||||
|
# not use this file except in compliance with the License. You may obtain
|
||||||
|
# a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||||
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||||
|
# License for the specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
|
||||||
|
from oslo_config import cfg
|
||||||
|
|
||||||
|
from watcher.common import clients
|
||||||
|
from watcher.common.metal_helper import ironic
|
||||||
|
from watcher.common.metal_helper import maas
|
||||||
|
|
||||||
|
CONF = cfg.CONF
|
||||||
|
|
||||||
|
|
||||||
|
def get_helper(osc=None):
|
||||||
|
# TODO(lpetrut): consider caching this client.
|
||||||
|
if not osc:
|
||||||
|
osc = clients.OpenStackClients()
|
||||||
|
|
||||||
|
if CONF.maas_client.url:
|
||||||
|
return maas.MaasHelper(osc)
|
||||||
|
else:
|
||||||
|
return ironic.IronicHelper(osc)
|
||||||
94
watcher/common/metal_helper/ironic.py
Normal file
94
watcher/common/metal_helper/ironic.py
Normal file
@@ -0,0 +1,94 @@
|
|||||||
|
# Copyright 2023 Cloudbase Solutions
|
||||||
|
# All Rights Reserved.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||||
|
# not use this file except in compliance with the License. You may obtain
|
||||||
|
# a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||||
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||||
|
# License for the specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
|
||||||
|
from oslo_log import log
|
||||||
|
|
||||||
|
from watcher.common.metal_helper import base
|
||||||
|
from watcher.common.metal_helper import constants as metal_constants
|
||||||
|
|
||||||
|
LOG = log.getLogger(__name__)
|
||||||
|
|
||||||
|
POWER_STATES_MAP = {
|
||||||
|
'power on': metal_constants.PowerState.ON,
|
||||||
|
'power off': metal_constants.PowerState.OFF,
|
||||||
|
# For now, we only use ON/OFF states
|
||||||
|
'rebooting': metal_constants.PowerState.ON,
|
||||||
|
'soft power off': metal_constants.PowerState.OFF,
|
||||||
|
'soft reboot': metal_constants.PowerState.ON,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class IronicNode(base.BaseMetalNode):
|
||||||
|
hv_up_when_powered_off = True
|
||||||
|
|
||||||
|
def __init__(self, ironic_node, nova_node, ironic_client):
|
||||||
|
super().__init__(nova_node)
|
||||||
|
|
||||||
|
self._ironic_client = ironic_client
|
||||||
|
self._ironic_node = ironic_node
|
||||||
|
|
||||||
|
def get_power_state(self):
|
||||||
|
return POWER_STATES_MAP.get(self._ironic_node.power_state,
|
||||||
|
metal_constants.PowerState.UNKNOWN)
|
||||||
|
|
||||||
|
def get_id(self):
|
||||||
|
return self._ironic_node.uuid
|
||||||
|
|
||||||
|
def power_on(self):
|
||||||
|
self._ironic_client.node.set_power_state(self.get_id(), "on")
|
||||||
|
|
||||||
|
def power_off(self):
|
||||||
|
self._ironic_client.node.set_power_state(self.get_id(), "off")
|
||||||
|
|
||||||
|
|
||||||
|
class IronicHelper(base.BaseMetalHelper):
|
||||||
|
@property
|
||||||
|
def _client(self):
|
||||||
|
if not getattr(self, "_cached_client", None):
|
||||||
|
self._cached_client = self._osc.ironic()
|
||||||
|
return self._cached_client
|
||||||
|
|
||||||
|
def list_compute_nodes(self):
|
||||||
|
out_list = []
|
||||||
|
# TODO(lpetrut): consider using "detailed=True" instead of making
|
||||||
|
# an additional GET request per node
|
||||||
|
node_list = self._client.node.list()
|
||||||
|
|
||||||
|
for node in node_list:
|
||||||
|
node_info = self._client.node.get(node.uuid)
|
||||||
|
hypervisor_id = node_info.extra.get('compute_node_id', None)
|
||||||
|
if hypervisor_id is None:
|
||||||
|
LOG.warning('Cannot find compute_node_id in extra '
|
||||||
|
'of ironic node %s', node.uuid)
|
||||||
|
continue
|
||||||
|
|
||||||
|
hypervisor_node = self.nova_client.hypervisors.get(hypervisor_id)
|
||||||
|
if hypervisor_node is None:
|
||||||
|
LOG.warning('Cannot find hypervisor %s', hypervisor_id)
|
||||||
|
continue
|
||||||
|
|
||||||
|
out_node = IronicNode(node, hypervisor_node, self._client)
|
||||||
|
out_list.append(out_node)
|
||||||
|
|
||||||
|
return out_list
|
||||||
|
|
||||||
|
def get_node(self, node_id):
|
||||||
|
ironic_node = self._client.node.get(node_id)
|
||||||
|
compute_node_id = ironic_node.extra.get('compute_node_id')
|
||||||
|
if compute_node_id:
|
||||||
|
compute_node = self.nova_client.hypervisors.get(compute_node_id)
|
||||||
|
else:
|
||||||
|
compute_node = None
|
||||||
|
return IronicNode(ironic_node, compute_node, self._client)
|
||||||
125
watcher/common/metal_helper/maas.py
Normal file
125
watcher/common/metal_helper/maas.py
Normal file
@@ -0,0 +1,125 @@
|
|||||||
|
# Copyright 2023 Cloudbase Solutions
|
||||||
|
# All Rights Reserved.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||||
|
# not use this file except in compliance with the License. You may obtain
|
||||||
|
# a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||||
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||||
|
# License for the specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
|
||||||
|
from oslo_config import cfg
|
||||||
|
from oslo_log import log
|
||||||
|
|
||||||
|
from watcher.common import exception
|
||||||
|
from watcher.common.metal_helper import base
|
||||||
|
from watcher.common.metal_helper import constants as metal_constants
|
||||||
|
from watcher.common import utils
|
||||||
|
|
||||||
|
CONF = cfg.CONF
|
||||||
|
LOG = log.getLogger(__name__)
|
||||||
|
|
||||||
|
try:
|
||||||
|
from maas.client import enum as maas_enum
|
||||||
|
except ImportError:
|
||||||
|
maas_enum = None
|
||||||
|
|
||||||
|
|
||||||
|
class MaasNode(base.BaseMetalNode):
|
||||||
|
hv_up_when_powered_off = False
|
||||||
|
|
||||||
|
def __init__(self, maas_node, nova_node, maas_client):
|
||||||
|
super().__init__(nova_node)
|
||||||
|
|
||||||
|
self._maas_client = maas_client
|
||||||
|
self._maas_node = maas_node
|
||||||
|
|
||||||
|
def get_power_state(self):
|
||||||
|
maas_state = utils.async_compat_call(
|
||||||
|
self._maas_node.query_power_state,
|
||||||
|
timeout=CONF.maas_client.timeout)
|
||||||
|
|
||||||
|
# python-libmaas may not be available, so we'll avoid a global
|
||||||
|
# variable.
|
||||||
|
power_states_map = {
|
||||||
|
maas_enum.PowerState.ON: metal_constants.PowerState.ON,
|
||||||
|
maas_enum.PowerState.OFF: metal_constants.PowerState.OFF,
|
||||||
|
maas_enum.PowerState.ERROR: metal_constants.PowerState.ERROR,
|
||||||
|
maas_enum.PowerState.UNKNOWN: metal_constants.PowerState.UNKNOWN,
|
||||||
|
}
|
||||||
|
return power_states_map.get(maas_state,
|
||||||
|
metal_constants.PowerState.UNKNOWN)
|
||||||
|
|
||||||
|
def get_id(self):
|
||||||
|
return self._maas_node.system_id
|
||||||
|
|
||||||
|
def power_on(self):
|
||||||
|
LOG.info("Powering on MAAS node: %s %s",
|
||||||
|
self._maas_node.fqdn,
|
||||||
|
self._maas_node.system_id)
|
||||||
|
utils.async_compat_call(
|
||||||
|
self._maas_node.power_on,
|
||||||
|
timeout=CONF.maas_client.timeout)
|
||||||
|
|
||||||
|
def power_off(self):
|
||||||
|
LOG.info("Powering off MAAS node: %s %s",
|
||||||
|
self._maas_node.fqdn,
|
||||||
|
self._maas_node.system_id)
|
||||||
|
utils.async_compat_call(
|
||||||
|
self._maas_node.power_off,
|
||||||
|
timeout=CONF.maas_client.timeout)
|
||||||
|
|
||||||
|
|
||||||
|
class MaasHelper(base.BaseMetalHelper):
|
||||||
|
def __init__(self, *args, **kwargs):
|
||||||
|
super().__init__(*args, **kwargs)
|
||||||
|
if not maas_enum:
|
||||||
|
raise exception.UnsupportedError(
|
||||||
|
"MAAS client unavailable. Please install python-libmaas.")
|
||||||
|
|
||||||
|
@property
|
||||||
|
def _client(self):
|
||||||
|
if not getattr(self, "_cached_client", None):
|
||||||
|
self._cached_client = self._osc.maas()
|
||||||
|
return self._cached_client
|
||||||
|
|
||||||
|
def list_compute_nodes(self):
|
||||||
|
out_list = []
|
||||||
|
node_list = utils.async_compat_call(
|
||||||
|
self._client.machines.list,
|
||||||
|
timeout=CONF.maas_client.timeout)
|
||||||
|
|
||||||
|
compute_nodes = self.nova_client.hypervisors.list()
|
||||||
|
compute_node_map = dict()
|
||||||
|
for compute_node in compute_nodes:
|
||||||
|
compute_node_map[compute_node.hypervisor_hostname] = compute_node
|
||||||
|
|
||||||
|
for node in node_list:
|
||||||
|
hypervisor_node = compute_node_map.get(node.fqdn)
|
||||||
|
if not hypervisor_node:
|
||||||
|
LOG.info('Cannot find hypervisor %s', node.fqdn)
|
||||||
|
continue
|
||||||
|
|
||||||
|
out_node = MaasNode(node, hypervisor_node, self._client)
|
||||||
|
out_list.append(out_node)
|
||||||
|
|
||||||
|
return out_list
|
||||||
|
|
||||||
|
def _get_compute_node_by_hostname(self, hostname):
|
||||||
|
compute_nodes = self.nova_client.hypervisors.search(
|
||||||
|
hostname, detailed=True)
|
||||||
|
for compute_node in compute_nodes:
|
||||||
|
if compute_node.hypervisor_hostname == hostname:
|
||||||
|
return compute_node
|
||||||
|
|
||||||
|
def get_node(self, node_id):
|
||||||
|
maas_node = utils.async_compat_call(
|
||||||
|
self._client.machines.get, node_id,
|
||||||
|
timeout=CONF.maas_client.timeout)
|
||||||
|
compute_node = self._get_compute_node_by_hostname(maas_node.fqdn)
|
||||||
|
return MaasNode(maas_node, compute_node, self._client)
|
||||||
@@ -11,6 +11,7 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
|
from http import HTTPStatus
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log as logging
|
from oslo_log import log as logging
|
||||||
|
|
||||||
@@ -53,7 +54,7 @@ class PlacementHelper(object):
|
|||||||
if rp_name:
|
if rp_name:
|
||||||
url += '?name=%s' % rp_name
|
url += '?name=%s' % rp_name
|
||||||
resp = self.get(url)
|
resp = self.get(url)
|
||||||
if resp.status_code == 200:
|
if resp.status_code == HTTPStatus.OK:
|
||||||
json_resp = resp.json()
|
json_resp = resp.json()
|
||||||
return json_resp['resource_providers']
|
return json_resp['resource_providers']
|
||||||
|
|
||||||
@@ -77,7 +78,7 @@ class PlacementHelper(object):
|
|||||||
"""
|
"""
|
||||||
url = '/resource_providers/%s/inventories' % rp_uuid
|
url = '/resource_providers/%s/inventories' % rp_uuid
|
||||||
resp = self.get(url)
|
resp = self.get(url)
|
||||||
if resp.status_code == 200:
|
if resp.status_code == HTTPStatus.OK:
|
||||||
json = resp.json()
|
json = resp.json()
|
||||||
return json['inventories']
|
return json['inventories']
|
||||||
msg = ("Failed to get resource provider %(rp_uuid)s inventories. "
|
msg = ("Failed to get resource provider %(rp_uuid)s inventories. "
|
||||||
@@ -97,7 +98,7 @@ class PlacementHelper(object):
|
|||||||
"""
|
"""
|
||||||
resp = self.get("/resource_providers/%s/traits" % rp_uuid)
|
resp = self.get("/resource_providers/%s/traits" % rp_uuid)
|
||||||
|
|
||||||
if resp.status_code == 200:
|
if resp.status_code == HTTPStatus.OK:
|
||||||
json = resp.json()
|
json = resp.json()
|
||||||
return json['traits']
|
return json['traits']
|
||||||
msg = ("Failed to get resource provider %(rp_uuid)s traits. "
|
msg = ("Failed to get resource provider %(rp_uuid)s traits. "
|
||||||
@@ -118,7 +119,7 @@ class PlacementHelper(object):
|
|||||||
"""
|
"""
|
||||||
url = '/allocations/%s' % consumer_uuid
|
url = '/allocations/%s' % consumer_uuid
|
||||||
resp = self.get(url)
|
resp = self.get(url)
|
||||||
if resp.status_code == 200:
|
if resp.status_code == HTTPStatus.OK:
|
||||||
json = resp.json()
|
json = resp.json()
|
||||||
return json['allocations']
|
return json['allocations']
|
||||||
msg = ("Failed to get allocations for consumer %(c_uuid). "
|
msg = ("Failed to get allocations for consumer %(c_uuid). "
|
||||||
@@ -139,7 +140,7 @@ class PlacementHelper(object):
|
|||||||
"""
|
"""
|
||||||
url = '/resource_providers/%s/usages' % rp_uuid
|
url = '/resource_providers/%s/usages' % rp_uuid
|
||||||
resp = self.get(url)
|
resp = self.get(url)
|
||||||
if resp.status_code == 200:
|
if resp.status_code == HTTPStatus.OK:
|
||||||
json = resp.json()
|
json = resp.json()
|
||||||
return json['usages']
|
return json['usages']
|
||||||
msg = ("Failed to get resource provider %(rp_uuid)s usages. "
|
msg = ("Failed to get resource provider %(rp_uuid)s usages. "
|
||||||
@@ -164,7 +165,7 @@ class PlacementHelper(object):
|
|||||||
"""
|
"""
|
||||||
url = "/allocation_candidates?%s" % resources
|
url = "/allocation_candidates?%s" % resources
|
||||||
resp = self.get(url)
|
resp = self.get(url)
|
||||||
if resp.status_code == 200:
|
if resp.status_code == HTTPStatus.OK:
|
||||||
data = resp.json()
|
data = resp.json()
|
||||||
return data['provider_summaries']
|
return data['provider_summaries']
|
||||||
|
|
||||||
|
|||||||
@@ -18,6 +18,7 @@
|
|||||||
import sys
|
import sys
|
||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
|
from oslo_policy import opts
|
||||||
from oslo_policy import policy
|
from oslo_policy import policy
|
||||||
|
|
||||||
from watcher.common import exception
|
from watcher.common import exception
|
||||||
@@ -26,6 +27,12 @@ from watcher.common import policies
|
|||||||
_ENFORCER = None
|
_ENFORCER = None
|
||||||
CONF = cfg.CONF
|
CONF = cfg.CONF
|
||||||
|
|
||||||
|
# TODO(gmann): Remove setting the default value of config policy_file
|
||||||
|
# once oslo_policy change the default value to 'policy.yaml'.
|
||||||
|
# https://github.com/openstack/oslo.policy/blob/a626ad12fe5a3abd49d70e3e5b95589d279ab578/oslo_policy/opts.py#L49
|
||||||
|
DEFAULT_POLICY_FILE = 'policy.yaml'
|
||||||
|
opts.set_defaults(CONF, DEFAULT_POLICY_FILE)
|
||||||
|
|
||||||
|
|
||||||
# we can get a policy enforcer by this init.
|
# we can get a policy enforcer by this init.
|
||||||
# oslo policy support change policy rule dynamically.
|
# oslo policy support change policy rule dynamically.
|
||||||
|
|||||||
@@ -121,22 +121,40 @@ class RequestContextSerializer(messaging.Serializer):
|
|||||||
def get_client(target, version_cap=None, serializer=None):
|
def get_client(target, version_cap=None, serializer=None):
|
||||||
assert TRANSPORT is not None
|
assert TRANSPORT is not None
|
||||||
serializer = RequestContextSerializer(serializer)
|
serializer = RequestContextSerializer(serializer)
|
||||||
return messaging.RPCClient(TRANSPORT,
|
return messaging.get_rpc_client(
|
||||||
target,
|
TRANSPORT,
|
||||||
version_cap=version_cap,
|
target,
|
||||||
serializer=serializer)
|
version_cap=version_cap,
|
||||||
|
serializer=serializer
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def get_server(target, endpoints, serializer=None):
|
def get_server(target, endpoints, serializer=None):
|
||||||
assert TRANSPORT is not None
|
assert TRANSPORT is not None
|
||||||
access_policy = dispatcher.DefaultRPCAccessPolicy
|
access_policy = dispatcher.DefaultRPCAccessPolicy
|
||||||
serializer = RequestContextSerializer(serializer)
|
serializer = RequestContextSerializer(serializer)
|
||||||
return messaging.get_rpc_server(TRANSPORT,
|
return messaging.get_rpc_server(
|
||||||
target,
|
TRANSPORT,
|
||||||
endpoints,
|
target,
|
||||||
executor='eventlet',
|
endpoints,
|
||||||
serializer=serializer,
|
executor='eventlet',
|
||||||
access_policy=access_policy)
|
serializer=serializer,
|
||||||
|
access_policy=access_policy
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def get_notification_listener(targets, endpoints, serializer=None, pool=None):
|
||||||
|
assert NOTIFICATION_TRANSPORT is not None
|
||||||
|
serializer = RequestContextSerializer(serializer)
|
||||||
|
return messaging.get_notification_listener(
|
||||||
|
NOTIFICATION_TRANSPORT,
|
||||||
|
targets,
|
||||||
|
endpoints,
|
||||||
|
allow_requeue=False,
|
||||||
|
executor='eventlet',
|
||||||
|
pool=pool,
|
||||||
|
serializer=serializer
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def get_notifier(publisher_id):
|
def get_notifier(publisher_id):
|
||||||
|
|||||||
@@ -21,14 +21,12 @@ from oslo_concurrency import processutils
|
|||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import _options
|
from oslo_log import _options
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
import oslo_messaging as om
|
import oslo_messaging as messaging
|
||||||
from oslo_reports import guru_meditation_report as gmr
|
from oslo_reports import guru_meditation_report as gmr
|
||||||
from oslo_reports import opts as gmr_opts
|
from oslo_reports import opts as gmr_opts
|
||||||
from oslo_service import service
|
from oslo_service import service
|
||||||
from oslo_service import wsgi
|
from oslo_service import wsgi
|
||||||
|
|
||||||
from oslo_messaging.rpc import dispatcher
|
|
||||||
|
|
||||||
from watcher._i18n import _
|
from watcher._i18n import _
|
||||||
from watcher.api import app
|
from watcher.api import app
|
||||||
from watcher.common import config
|
from watcher.common import config
|
||||||
@@ -183,11 +181,6 @@ class Service(service.ServiceBase):
|
|||||||
]
|
]
|
||||||
self.notification_endpoints = self.manager.notification_endpoints
|
self.notification_endpoints = self.manager.notification_endpoints
|
||||||
|
|
||||||
self.serializer = rpc.RequestContextSerializer(
|
|
||||||
base.WatcherObjectSerializer())
|
|
||||||
|
|
||||||
self._transport = None
|
|
||||||
self._notification_transport = None
|
|
||||||
self._conductor_client = None
|
self._conductor_client = None
|
||||||
|
|
||||||
self.conductor_topic_handler = None
|
self.conductor_topic_handler = None
|
||||||
@@ -201,27 +194,17 @@ class Service(service.ServiceBase):
|
|||||||
self.notification_topics, self.notification_endpoints
|
self.notification_topics, self.notification_endpoints
|
||||||
)
|
)
|
||||||
|
|
||||||
@property
|
|
||||||
def transport(self):
|
|
||||||
if self._transport is None:
|
|
||||||
self._transport = om.get_rpc_transport(CONF)
|
|
||||||
return self._transport
|
|
||||||
|
|
||||||
@property
|
|
||||||
def notification_transport(self):
|
|
||||||
if self._notification_transport is None:
|
|
||||||
self._notification_transport = om.get_notification_transport(CONF)
|
|
||||||
return self._notification_transport
|
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def conductor_client(self):
|
def conductor_client(self):
|
||||||
if self._conductor_client is None:
|
if self._conductor_client is None:
|
||||||
target = om.Target(
|
target = messaging.Target(
|
||||||
topic=self.conductor_topic,
|
topic=self.conductor_topic,
|
||||||
version=self.API_VERSION,
|
version=self.API_VERSION,
|
||||||
)
|
)
|
||||||
self._conductor_client = om.RPCClient(
|
self._conductor_client = rpc.get_client(
|
||||||
self.transport, target, serializer=self.serializer)
|
target,
|
||||||
|
serializer=base.WatcherObjectSerializer()
|
||||||
|
)
|
||||||
return self._conductor_client
|
return self._conductor_client
|
||||||
|
|
||||||
@conductor_client.setter
|
@conductor_client.setter
|
||||||
@@ -229,21 +212,18 @@ class Service(service.ServiceBase):
|
|||||||
self.conductor_client = c
|
self.conductor_client = c
|
||||||
|
|
||||||
def build_topic_handler(self, topic_name, endpoints=()):
|
def build_topic_handler(self, topic_name, endpoints=()):
|
||||||
access_policy = dispatcher.DefaultRPCAccessPolicy
|
target = messaging.Target(
|
||||||
serializer = rpc.RequestContextSerializer(rpc.JsonPayloadSerializer())
|
|
||||||
target = om.Target(
|
|
||||||
topic=topic_name,
|
topic=topic_name,
|
||||||
# For compatibility, we can override it with 'host' opt
|
# For compatibility, we can override it with 'host' opt
|
||||||
server=CONF.host or socket.gethostname(),
|
server=CONF.host or socket.gethostname(),
|
||||||
version=self.api_version,
|
version=self.api_version,
|
||||||
)
|
)
|
||||||
return om.get_rpc_server(
|
return rpc.get_server(
|
||||||
self.transport, target, endpoints,
|
target, endpoints,
|
||||||
executor='eventlet', serializer=serializer,
|
serializer=rpc.JsonPayloadSerializer()
|
||||||
access_policy=access_policy)
|
)
|
||||||
|
|
||||||
def build_notification_handler(self, topic_names, endpoints=()):
|
def build_notification_handler(self, topic_names, endpoints=()):
|
||||||
serializer = rpc.RequestContextSerializer(rpc.JsonPayloadSerializer())
|
|
||||||
targets = []
|
targets = []
|
||||||
for topic in topic_names:
|
for topic in topic_names:
|
||||||
kwargs = {}
|
kwargs = {}
|
||||||
@@ -251,11 +231,13 @@ class Service(service.ServiceBase):
|
|||||||
exchange, topic = topic.split('.')
|
exchange, topic = topic.split('.')
|
||||||
kwargs['exchange'] = exchange
|
kwargs['exchange'] = exchange
|
||||||
kwargs['topic'] = topic
|
kwargs['topic'] = topic
|
||||||
targets.append(om.Target(**kwargs))
|
targets.append(messaging.Target(**kwargs))
|
||||||
return om.get_notification_listener(
|
|
||||||
self.notification_transport, targets, endpoints,
|
return rpc.get_notification_listener(
|
||||||
executor='eventlet', serializer=serializer,
|
targets, endpoints,
|
||||||
allow_requeue=False, pool=CONF.host)
|
serializer=rpc.JsonPayloadSerializer(),
|
||||||
|
pool=CONF.host
|
||||||
|
)
|
||||||
|
|
||||||
def start(self):
|
def start(self):
|
||||||
LOG.debug("Connecting to '%s'", CONF.transport_url)
|
LOG.debug("Connecting to '%s'", CONF.transport_url)
|
||||||
|
|||||||
@@ -15,11 +15,9 @@
|
|||||||
# under the License.
|
# under the License.
|
||||||
|
|
||||||
import abc
|
import abc
|
||||||
import six
|
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class ServiceManager(object, metaclass=abc.ABCMeta):
|
||||||
class ServiceManager(object):
|
|
||||||
|
|
||||||
@abc.abstractproperty
|
@abc.abstractproperty
|
||||||
def service_name(self):
|
def service_name(self):
|
||||||
|
|||||||
@@ -16,19 +16,22 @@
|
|||||||
|
|
||||||
"""Utilities and helper functions."""
|
"""Utilities and helper functions."""
|
||||||
|
|
||||||
|
import asyncio
|
||||||
import datetime
|
import datetime
|
||||||
|
import inspect
|
||||||
import random
|
import random
|
||||||
import re
|
import re
|
||||||
import string
|
import string
|
||||||
|
|
||||||
from croniter import croniter
|
from croniter import croniter
|
||||||
|
import eventlet
|
||||||
|
from eventlet import tpool
|
||||||
|
|
||||||
from jsonschema import validators
|
from jsonschema import validators
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
from oslo_utils import strutils
|
from oslo_utils import strutils
|
||||||
from oslo_utils import uuidutils
|
from oslo_utils import uuidutils
|
||||||
import six
|
|
||||||
|
|
||||||
from watcher.common import exception
|
from watcher.common import exception
|
||||||
|
|
||||||
@@ -82,7 +85,7 @@ def safe_rstrip(value, chars=None):
|
|||||||
:return: Stripped value.
|
:return: Stripped value.
|
||||||
|
|
||||||
"""
|
"""
|
||||||
if not isinstance(value, six.string_types):
|
if not isinstance(value, str):
|
||||||
LOG.warning(
|
LOG.warning(
|
||||||
"Failed to remove trailing character. Returning original object."
|
"Failed to remove trailing character. Returning original object."
|
||||||
"Supplied object is not a string: %s,", value)
|
"Supplied object is not a string: %s,", value)
|
||||||
@@ -104,7 +107,7 @@ def is_hostname_safe(hostname):
|
|||||||
|
|
||||||
"""
|
"""
|
||||||
m = r'^[a-z0-9]([a-z0-9\-]{0,61}[a-z0-9])?$'
|
m = r'^[a-z0-9]([a-z0-9\-]{0,61}[a-z0-9])?$'
|
||||||
return (isinstance(hostname, six.string_types) and
|
return (isinstance(hostname, str) and
|
||||||
(re.match(m, hostname) is not None))
|
(re.match(m, hostname) is not None))
|
||||||
|
|
||||||
|
|
||||||
@@ -163,3 +166,37 @@ Draft4Validator = validators.Draft4Validator
|
|||||||
def random_string(n):
|
def random_string(n):
|
||||||
return ''.join([random.choice(
|
return ''.join([random.choice(
|
||||||
string.ascii_letters + string.digits) for i in range(n)])
|
string.ascii_letters + string.digits) for i in range(n)])
|
||||||
|
|
||||||
|
|
||||||
|
# Some clients (e.g. MAAS) use asyncio, which isn't compatible with Eventlet.
|
||||||
|
# As a workaround, we're delegating such calls to a native thread.
|
||||||
|
def async_compat_call(f, *args, **kwargs):
|
||||||
|
timeout = kwargs.pop('timeout', None)
|
||||||
|
|
||||||
|
async def async_wrapper():
|
||||||
|
ret = f(*args, **kwargs)
|
||||||
|
if inspect.isawaitable(ret):
|
||||||
|
return await asyncio.wait_for(ret, timeout)
|
||||||
|
return ret
|
||||||
|
|
||||||
|
def tpool_wrapper():
|
||||||
|
# This will run in a separate native thread. Ideally, there should be
|
||||||
|
# a single thread permanently running an asyncio loop, but for
|
||||||
|
# convenience we'll use eventlet.tpool, which leverages a thread pool.
|
||||||
|
#
|
||||||
|
# That being considered, we're setting up a temporary asyncio loop to
|
||||||
|
# handle this call.
|
||||||
|
loop = asyncio.new_event_loop()
|
||||||
|
try:
|
||||||
|
asyncio.set_event_loop(loop)
|
||||||
|
return loop.run_until_complete(async_wrapper())
|
||||||
|
finally:
|
||||||
|
loop.close()
|
||||||
|
|
||||||
|
# We'll use eventlet timeouts as an extra precaution and asyncio timeouts
|
||||||
|
# to avoid lingering threads. For consistency, we'll convert eventlet
|
||||||
|
# timeout exceptions to asyncio timeout errors.
|
||||||
|
with eventlet.timeout.Timeout(
|
||||||
|
seconds=timeout,
|
||||||
|
exception=asyncio.TimeoutError("Timeout: %ss" % timeout)):
|
||||||
|
return tpool.execute(tpool_wrapper)
|
||||||
|
|||||||
@@ -35,6 +35,7 @@ from watcher.conf import grafana_client
|
|||||||
from watcher.conf import grafana_translators
|
from watcher.conf import grafana_translators
|
||||||
from watcher.conf import ironic_client
|
from watcher.conf import ironic_client
|
||||||
from watcher.conf import keystone_client
|
from watcher.conf import keystone_client
|
||||||
|
from watcher.conf import maas_client
|
||||||
from watcher.conf import monasca_client
|
from watcher.conf import monasca_client
|
||||||
from watcher.conf import neutron_client
|
from watcher.conf import neutron_client
|
||||||
from watcher.conf import nova_client
|
from watcher.conf import nova_client
|
||||||
@@ -54,6 +55,7 @@ db.register_opts(CONF)
|
|||||||
planner.register_opts(CONF)
|
planner.register_opts(CONF)
|
||||||
applier.register_opts(CONF)
|
applier.register_opts(CONF)
|
||||||
decision_engine.register_opts(CONF)
|
decision_engine.register_opts(CONF)
|
||||||
|
maas_client.register_opts(CONF)
|
||||||
monasca_client.register_opts(CONF)
|
monasca_client.register_opts(CONF)
|
||||||
nova_client.register_opts(CONF)
|
nova_client.register_opts(CONF)
|
||||||
glance_client.register_opts(CONF)
|
glance_client.register_opts(CONF)
|
||||||
|
|||||||
@@ -43,11 +43,20 @@ APPLIER_MANAGER_OPTS = [
|
|||||||
help='Select the engine to use to execute the workflow'),
|
help='Select the engine to use to execute the workflow'),
|
||||||
]
|
]
|
||||||
|
|
||||||
|
APPLIER_OPTS = [
|
||||||
|
cfg.BoolOpt('rollback_when_actionplan_failed',
|
||||||
|
default=False,
|
||||||
|
help='If set True, the failed actionplan will rollback '
|
||||||
|
'when executing. Defaule value is False.'),
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
def register_opts(conf):
|
def register_opts(conf):
|
||||||
conf.register_group(watcher_applier)
|
conf.register_group(watcher_applier)
|
||||||
conf.register_opts(APPLIER_MANAGER_OPTS, group=watcher_applier)
|
conf.register_opts(APPLIER_MANAGER_OPTS, group=watcher_applier)
|
||||||
|
conf.register_opts(APPLIER_OPTS, group=watcher_applier)
|
||||||
|
|
||||||
|
|
||||||
def list_opts():
|
def list_opts():
|
||||||
return [(watcher_applier, APPLIER_MANAGER_OPTS)]
|
return [(watcher_applier, APPLIER_MANAGER_OPTS),
|
||||||
|
(watcher_applier, APPLIER_OPTS)]
|
||||||
|
|||||||
@@ -134,7 +134,13 @@ GRAFANA_CLIENT_OPTS = [
|
|||||||
"InfluxDB this will be the retention period. "
|
"InfluxDB this will be the retention period. "
|
||||||
"These queries will need to be constructed using tools "
|
"These queries will need to be constructed using tools "
|
||||||
"such as Postman. Example: SELECT cpu FROM {4}."
|
"such as Postman. Example: SELECT cpu FROM {4}."
|
||||||
"cpu_percent WHERE host == '{1}' AND time > now()-{2}s")]
|
"cpu_percent WHERE host == '{1}' AND time > now()-{2}s"),
|
||||||
|
cfg.IntOpt('http_timeout',
|
||||||
|
min=0,
|
||||||
|
default=60,
|
||||||
|
mutable=True,
|
||||||
|
help='Timeout for Grafana request')
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
def register_opts(conf):
|
def register_opts(conf):
|
||||||
|
|||||||
38
watcher/conf/maas_client.py
Normal file
38
watcher/conf/maas_client.py
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
# Copyright 2023 Cloudbase Solutions
|
||||||
|
# All Rights Reserved.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||||
|
# not use this file except in compliance with the License. You may obtain
|
||||||
|
# a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||||
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||||
|
# License for the specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
|
||||||
|
|
||||||
|
from oslo_config import cfg
|
||||||
|
|
||||||
|
maas_client = cfg.OptGroup(name='maas_client',
|
||||||
|
title='Configuration Options for MaaS')
|
||||||
|
|
||||||
|
MAAS_CLIENT_OPTS = [
|
||||||
|
cfg.StrOpt('url',
|
||||||
|
help='MaaS URL, example: http://1.2.3.4:5240/MAAS'),
|
||||||
|
cfg.StrOpt('api_key',
|
||||||
|
help='MaaS API authentication key.'),
|
||||||
|
cfg.IntOpt('timeout',
|
||||||
|
default=60,
|
||||||
|
help='MaaS client operation timeout in seconds.')]
|
||||||
|
|
||||||
|
|
||||||
|
def register_opts(conf):
|
||||||
|
conf.register_group(maas_client)
|
||||||
|
conf.register_opts(MAAS_CLIENT_OPTS, group=maas_client)
|
||||||
|
|
||||||
|
|
||||||
|
def list_opts():
|
||||||
|
return [(maas_client, MAAS_CLIENT_OPTS)]
|
||||||
@@ -18,7 +18,6 @@ Base classes for storage engines
|
|||||||
import abc
|
import abc
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_db import api as db_api
|
from oslo_db import api as db_api
|
||||||
import six
|
|
||||||
|
|
||||||
_BACKEND_MAPPING = {'sqlalchemy': 'watcher.db.sqlalchemy.api'}
|
_BACKEND_MAPPING = {'sqlalchemy': 'watcher.db.sqlalchemy.api'}
|
||||||
IMPL = db_api.DBAPI.from_config(cfg.CONF, backend_mapping=_BACKEND_MAPPING,
|
IMPL = db_api.DBAPI.from_config(cfg.CONF, backend_mapping=_BACKEND_MAPPING,
|
||||||
@@ -30,8 +29,7 @@ def get_instance():
|
|||||||
return IMPL
|
return IMPL
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class BaseConnection(object, metaclass=abc.ABCMeta):
|
||||||
class BaseConnection(object):
|
|
||||||
"""Base class for storage system connections."""
|
"""Base class for storage system connections."""
|
||||||
|
|
||||||
@abc.abstractmethod
|
@abc.abstractmethod
|
||||||
|
|||||||
@@ -15,8 +15,6 @@
|
|||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
#
|
#
|
||||||
|
|
||||||
from __future__ import print_function
|
|
||||||
|
|
||||||
import collections
|
import collections
|
||||||
import datetime
|
import datetime
|
||||||
import itertools
|
import itertools
|
||||||
@@ -25,7 +23,6 @@ import sys
|
|||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
from oslo_utils import strutils
|
from oslo_utils import strutils
|
||||||
import prettytable as ptable
|
import prettytable as ptable
|
||||||
from six.moves import input
|
|
||||||
|
|
||||||
from watcher._i18n import _
|
from watcher._i18n import _
|
||||||
from watcher._i18n import lazy_translation_enabled
|
from watcher._i18n import lazy_translation_enabled
|
||||||
|
|||||||
@@ -6,6 +6,7 @@ Create Date: 2017-03-24 11:21:29.036532
|
|||||||
|
|
||||||
"""
|
"""
|
||||||
from alembic import op
|
from alembic import op
|
||||||
|
from sqlalchemy import inspect
|
||||||
import sqlalchemy as sa
|
import sqlalchemy as sa
|
||||||
|
|
||||||
from watcher.db.sqlalchemy import models
|
from watcher.db.sqlalchemy import models
|
||||||
@@ -14,8 +15,17 @@ from watcher.db.sqlalchemy import models
|
|||||||
revision = '0f6042416884'
|
revision = '0f6042416884'
|
||||||
down_revision = '001'
|
down_revision = '001'
|
||||||
|
|
||||||
|
def _table_exists(table_name):
|
||||||
|
bind = op.get_context().bind
|
||||||
|
insp = inspect(bind)
|
||||||
|
names = insp.get_table_names()
|
||||||
|
return any(t == table_name for t in names)
|
||||||
|
|
||||||
|
|
||||||
def upgrade():
|
def upgrade():
|
||||||
|
if _table_exists('apscheduler_jobs'):
|
||||||
|
return
|
||||||
|
|
||||||
op.create_table(
|
op.create_table(
|
||||||
'apscheduler_jobs',
|
'apscheduler_jobs',
|
||||||
sa.Column('id', sa.Unicode(191, _warn_on_bytestring=False),
|
sa.Column('id', sa.Unicode(191, _warn_on_bytestring=False),
|
||||||
|
|||||||
@@ -248,26 +248,31 @@ class Connection(api.BaseConnection):
|
|||||||
return query
|
return query
|
||||||
|
|
||||||
def _create(self, model, values):
|
def _create(self, model, values):
|
||||||
obj = model()
|
session = get_session()
|
||||||
cleaned_values = {k: v for k, v in values.items()
|
with session.begin():
|
||||||
if k not in self._get_relationships(model)}
|
obj = model()
|
||||||
obj.update(cleaned_values)
|
cleaned_values = {k: v for k, v in values.items()
|
||||||
obj.save()
|
if k not in self._get_relationships(model)}
|
||||||
|
obj.update(cleaned_values)
|
||||||
|
obj.save(session=session)
|
||||||
|
session.commit()
|
||||||
return obj
|
return obj
|
||||||
|
|
||||||
def _get(self, context, model, fieldname, value, eager):
|
def _get(self, context, model, fieldname, value, eager):
|
||||||
query = model_query(model)
|
session = get_session()
|
||||||
if eager:
|
with session.begin():
|
||||||
query = self._set_eager_options(model, query)
|
query = model_query(model, session=session)
|
||||||
|
if eager:
|
||||||
|
query = self._set_eager_options(model, query)
|
||||||
|
|
||||||
query = query.filter(getattr(model, fieldname) == value)
|
query = query.filter(getattr(model, fieldname) == value)
|
||||||
if not context.show_deleted:
|
if not context.show_deleted:
|
||||||
query = query.filter(model.deleted_at.is_(None))
|
query = query.filter(model.deleted_at.is_(None))
|
||||||
|
|
||||||
try:
|
try:
|
||||||
obj = query.one()
|
obj = query.one()
|
||||||
except exc.NoResultFound:
|
except exc.NoResultFound:
|
||||||
raise exception.ResourceNotFound(name=model.__name__, id=value)
|
raise exception.ResourceNotFound(name=model.__name__, id=value)
|
||||||
|
|
||||||
return obj
|
return obj
|
||||||
|
|
||||||
@@ -278,7 +283,7 @@ class Connection(api.BaseConnection):
|
|||||||
query = model_query(model, session=session)
|
query = model_query(model, session=session)
|
||||||
query = add_identity_filter(query, id_)
|
query = add_identity_filter(query, id_)
|
||||||
try:
|
try:
|
||||||
ref = query.with_lockmode('update').one()
|
ref = query.with_for_update().one()
|
||||||
except exc.NoResultFound:
|
except exc.NoResultFound:
|
||||||
raise exception.ResourceNotFound(name=model.__name__, id=id_)
|
raise exception.ResourceNotFound(name=model.__name__, id=id_)
|
||||||
|
|
||||||
@@ -815,7 +820,7 @@ class Connection(api.BaseConnection):
|
|||||||
query = model_query(models.Action, session=session)
|
query = model_query(models.Action, session=session)
|
||||||
query = add_identity_filter(query, action_id)
|
query = add_identity_filter(query, action_id)
|
||||||
try:
|
try:
|
||||||
ref = query.with_lockmode('update').one()
|
ref = query.with_for_update().one()
|
||||||
except exc.NoResultFound:
|
except exc.NoResultFound:
|
||||||
raise exception.ActionNotFound(action=action_id)
|
raise exception.ActionNotFound(action=action_id)
|
||||||
|
|
||||||
@@ -900,7 +905,7 @@ class Connection(api.BaseConnection):
|
|||||||
query = model_query(models.ActionPlan, session=session)
|
query = model_query(models.ActionPlan, session=session)
|
||||||
query = add_identity_filter(query, action_plan_id)
|
query = add_identity_filter(query, action_plan_id)
|
||||||
try:
|
try:
|
||||||
ref = query.with_lockmode('update').one()
|
ref = query.with_for_update().one()
|
||||||
except exc.NoResultFound:
|
except exc.NoResultFound:
|
||||||
raise exception.ActionPlanNotFound(action_plan=action_plan_id)
|
raise exception.ActionPlanNotFound(action_plan=action_plan_id)
|
||||||
|
|
||||||
|
|||||||
@@ -18,7 +18,6 @@ SQLAlchemy models for watcher service
|
|||||||
|
|
||||||
from oslo_db.sqlalchemy import models
|
from oslo_db.sqlalchemy import models
|
||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
import six.moves.urllib.parse as urlparse
|
|
||||||
from sqlalchemy import Boolean
|
from sqlalchemy import Boolean
|
||||||
from sqlalchemy import Column
|
from sqlalchemy import Column
|
||||||
from sqlalchemy import DateTime
|
from sqlalchemy import DateTime
|
||||||
@@ -33,7 +32,7 @@ from sqlalchemy import String
|
|||||||
from sqlalchemy import Text
|
from sqlalchemy import Text
|
||||||
from sqlalchemy.types import TypeDecorator, TEXT
|
from sqlalchemy.types import TypeDecorator, TEXT
|
||||||
from sqlalchemy import UniqueConstraint
|
from sqlalchemy import UniqueConstraint
|
||||||
|
import urllib.parse as urlparse
|
||||||
from watcher import conf
|
from watcher import conf
|
||||||
|
|
||||||
CONF = conf.CONF
|
CONF = conf.CONF
|
||||||
|
|||||||
@@ -18,7 +18,6 @@
|
|||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
#
|
#
|
||||||
import abc
|
import abc
|
||||||
import six
|
|
||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
@@ -36,9 +35,11 @@ CONF = cfg.CONF
|
|||||||
LOG = log.getLogger(__name__)
|
LOG = log.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class BaseMetaClass(service.Singleton, abc.ABCMeta):
|
||||||
@six.add_metaclass(service.Singleton)
|
pass
|
||||||
class BaseAuditHandler(object):
|
|
||||||
|
|
||||||
|
class BaseAuditHandler(object, metaclass=BaseMetaClass):
|
||||||
|
|
||||||
@abc.abstractmethod
|
@abc.abstractmethod
|
||||||
def execute(self, audit, request_context):
|
def execute(self, audit, request_context):
|
||||||
@@ -57,8 +58,7 @@ class BaseAuditHandler(object):
|
|||||||
raise NotImplementedError()
|
raise NotImplementedError()
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class AuditHandler(BaseAuditHandler, metaclass=abc.ABCMeta):
|
||||||
class AuditHandler(BaseAuditHandler):
|
|
||||||
|
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
super(AuditHandler, self).__init__()
|
super(AuditHandler, self).__init__()
|
||||||
|
|||||||
@@ -19,6 +19,8 @@ import time
|
|||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
|
|
||||||
|
from watcher.common import exception
|
||||||
|
|
||||||
CONF = cfg.CONF
|
CONF = cfg.CONF
|
||||||
LOG = log.getLogger(__name__)
|
LOG = log.getLogger(__name__)
|
||||||
|
|
||||||
@@ -54,7 +56,14 @@ class DataSourceBase(object):
|
|||||||
instance_root_disk_size=None,
|
instance_root_disk_size=None,
|
||||||
)
|
)
|
||||||
|
|
||||||
def query_retry(self, f, *args, **kwargs):
|
def _get_meter(self, meter_name):
|
||||||
|
"""Retrieve the meter from the metric map or raise error"""
|
||||||
|
meter = self.METRIC_MAP.get(meter_name)
|
||||||
|
if meter is None:
|
||||||
|
raise exception.MetricNotAvailable(metric=meter_name)
|
||||||
|
return meter
|
||||||
|
|
||||||
|
def query_retry(self, f, *args, ignored_exc=None, **kwargs):
|
||||||
"""Attempts to retrieve metrics from the external service
|
"""Attempts to retrieve metrics from the external service
|
||||||
|
|
||||||
Attempts to access data from the external service and handles
|
Attempts to access data from the external service and handles
|
||||||
@@ -62,15 +71,23 @@ class DataSourceBase(object):
|
|||||||
to the value of query_max_retries
|
to the value of query_max_retries
|
||||||
:param f: The method that performs the actual querying for metrics
|
:param f: The method that performs the actual querying for metrics
|
||||||
:param args: Array of arguments supplied to the method
|
:param args: Array of arguments supplied to the method
|
||||||
|
:param ignored_exc: An exception or tuple of exceptions that shouldn't
|
||||||
|
be retried, for example "NotFound" exceptions.
|
||||||
:param kwargs: The amount of arguments supplied to the method
|
:param kwargs: The amount of arguments supplied to the method
|
||||||
:return: The value as retrieved from the external service
|
:return: The value as retrieved from the external service
|
||||||
"""
|
"""
|
||||||
|
|
||||||
num_retries = CONF.watcher_datasources.query_max_retries
|
num_retries = CONF.watcher_datasources.query_max_retries
|
||||||
timeout = CONF.watcher_datasources.query_timeout
|
timeout = CONF.watcher_datasources.query_timeout
|
||||||
|
ignored_exc = ignored_exc or tuple()
|
||||||
|
|
||||||
for i in range(num_retries):
|
for i in range(num_retries):
|
||||||
try:
|
try:
|
||||||
return f(*args, **kwargs)
|
return f(*args, **kwargs)
|
||||||
|
except ignored_exc as e:
|
||||||
|
LOG.debug("Got an ignored exception (%s) while calling: %s ",
|
||||||
|
e, f)
|
||||||
|
return
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
LOG.exception(e)
|
LOG.exception(e)
|
||||||
self.query_retry_reset(e)
|
self.query_retry_reset(e)
|
||||||
@@ -122,6 +139,30 @@ class DataSourceBase(object):
|
|||||||
|
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
@abc.abstractmethod
|
||||||
|
def statistic_series(self, resource=None, resource_type=None,
|
||||||
|
meter_name=None, start_time=None, end_time=None,
|
||||||
|
granularity=300):
|
||||||
|
"""Retrieves metrics based on the specified parameters over a period
|
||||||
|
|
||||||
|
:param resource: Resource object as defined in watcher models such as
|
||||||
|
ComputeNode and Instance
|
||||||
|
:param resource_type: Indicates which type of object is supplied
|
||||||
|
to the resource parameter
|
||||||
|
:param meter_name: The desired metric to retrieve as key from
|
||||||
|
METRIC_MAP
|
||||||
|
:param start_time: The datetime to start retrieving metrics for
|
||||||
|
:type start_time: datetime.datetime
|
||||||
|
:param end_time: The datetime to limit the retrieval of metrics to
|
||||||
|
:type end_time: datetime.datetime
|
||||||
|
:param granularity: Interval between samples in measurements in
|
||||||
|
seconds
|
||||||
|
:return: Dictionary of key value pairs with timestamps and metric
|
||||||
|
values
|
||||||
|
"""
|
||||||
|
|
||||||
|
pass
|
||||||
|
|
||||||
@abc.abstractmethod
|
@abc.abstractmethod
|
||||||
def get_host_cpu_usage(self, resource, period, aggregate,
|
def get_host_cpu_usage(self, resource, period, aggregate,
|
||||||
granularity=None):
|
granularity=None):
|
||||||
|
|||||||
@@ -161,9 +161,7 @@ class CeilometerHelper(base.DataSourceBase):
|
|||||||
end_time = datetime.datetime.utcnow()
|
end_time = datetime.datetime.utcnow()
|
||||||
start_time = end_time - datetime.timedelta(seconds=int(period))
|
start_time = end_time - datetime.timedelta(seconds=int(period))
|
||||||
|
|
||||||
meter = self.METRIC_MAP.get(meter_name)
|
meter = self._get_meter(meter_name)
|
||||||
if meter is None:
|
|
||||||
raise exception.MetricNotAvailable(metric=meter_name)
|
|
||||||
|
|
||||||
if aggregate == 'mean':
|
if aggregate == 'mean':
|
||||||
aggregate = 'avg'
|
aggregate = 'avg'
|
||||||
@@ -194,6 +192,12 @@ class CeilometerHelper(base.DataSourceBase):
|
|||||||
item_value *= 10
|
item_value *= 10
|
||||||
return item_value
|
return item_value
|
||||||
|
|
||||||
|
def statistic_series(self, resource=None, resource_type=None,
|
||||||
|
meter_name=None, start_time=None, end_time=None,
|
||||||
|
granularity=300):
|
||||||
|
raise NotImplementedError(
|
||||||
|
_('Ceilometer helper does not support statistic series method'))
|
||||||
|
|
||||||
def get_host_cpu_usage(self, resource, period,
|
def get_host_cpu_usage(self, resource, period,
|
||||||
aggregate, granularity=None):
|
aggregate, granularity=None):
|
||||||
|
|
||||||
|
|||||||
@@ -19,11 +19,11 @@
|
|||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
from datetime import timedelta
|
from datetime import timedelta
|
||||||
|
|
||||||
|
from gnocchiclient import exceptions as gnc_exc
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
|
|
||||||
from watcher.common import clients
|
from watcher.common import clients
|
||||||
from watcher.common import exception
|
|
||||||
from watcher.decision_engine.datasources import base
|
from watcher.decision_engine.datasources import base
|
||||||
|
|
||||||
CONF = cfg.CONF
|
CONF = cfg.CONF
|
||||||
@@ -39,7 +39,7 @@ class GnocchiHelper(base.DataSourceBase):
|
|||||||
host_inlet_temp='hardware.ipmi.node.temperature',
|
host_inlet_temp='hardware.ipmi.node.temperature',
|
||||||
host_airflow='hardware.ipmi.node.airflow',
|
host_airflow='hardware.ipmi.node.airflow',
|
||||||
host_power='hardware.ipmi.node.power',
|
host_power='hardware.ipmi.node.power',
|
||||||
instance_cpu_usage='cpu_util',
|
instance_cpu_usage='cpu',
|
||||||
instance_ram_usage='memory.resident',
|
instance_ram_usage='memory.resident',
|
||||||
instance_ram_allocated='memory',
|
instance_ram_allocated='memory',
|
||||||
instance_l3_cache_usage='cpu_l3_cache',
|
instance_l3_cache_usage='cpu_l3_cache',
|
||||||
@@ -72,9 +72,7 @@ class GnocchiHelper(base.DataSourceBase):
|
|||||||
stop_time = datetime.utcnow()
|
stop_time = datetime.utcnow()
|
||||||
start_time = stop_time - timedelta(seconds=(int(period)))
|
start_time = stop_time - timedelta(seconds=(int(period)))
|
||||||
|
|
||||||
meter = self.METRIC_MAP.get(meter_name)
|
meter = self._get_meter(meter_name)
|
||||||
if meter is None:
|
|
||||||
raise exception.MetricNotAvailable(metric=meter_name)
|
|
||||||
|
|
||||||
if aggregate == 'count':
|
if aggregate == 'count':
|
||||||
aggregate = 'mean'
|
aggregate = 'mean'
|
||||||
@@ -87,7 +85,9 @@ class GnocchiHelper(base.DataSourceBase):
|
|||||||
kwargs = dict(query={"=": {"original_resource_id": resource_id}},
|
kwargs = dict(query={"=": {"original_resource_id": resource_id}},
|
||||||
limit=1)
|
limit=1)
|
||||||
resources = self.query_retry(
|
resources = self.query_retry(
|
||||||
f=self.gnocchi.resource.search, **kwargs)
|
f=self.gnocchi.resource.search,
|
||||||
|
ignored_exc=gnc_exc.NotFound,
|
||||||
|
**kwargs)
|
||||||
|
|
||||||
if not resources:
|
if not resources:
|
||||||
LOG.warning("The {0} resource {1} could not be "
|
LOG.warning("The {0} resource {1} could not be "
|
||||||
@@ -96,6 +96,25 @@ class GnocchiHelper(base.DataSourceBase):
|
|||||||
|
|
||||||
resource_id = resources[0]['id']
|
resource_id = resources[0]['id']
|
||||||
|
|
||||||
|
if meter_name == "instance_cpu_usage":
|
||||||
|
if resource_type != "instance":
|
||||||
|
LOG.warning("Unsupported resource type for metric "
|
||||||
|
"'instance_cpu_usage': ", resource_type)
|
||||||
|
return
|
||||||
|
|
||||||
|
# The "cpu_util" gauge (percentage) metric has been removed.
|
||||||
|
# We're going to obtain the same result by using the rate of change
|
||||||
|
# aggregate operation.
|
||||||
|
if aggregate not in ("mean", "rate:mean"):
|
||||||
|
LOG.warning("Unsupported aggregate for instance_cpu_usage "
|
||||||
|
"metric: %s. "
|
||||||
|
"Supported aggregates: mean, rate:mean ",
|
||||||
|
aggregate)
|
||||||
|
return
|
||||||
|
|
||||||
|
# TODO(lpetrut): consider supporting other aggregates.
|
||||||
|
aggregate = "rate:mean"
|
||||||
|
|
||||||
raw_kwargs = dict(
|
raw_kwargs = dict(
|
||||||
metric=meter,
|
metric=meter,
|
||||||
start=start_time,
|
start=start_time,
|
||||||
@@ -108,7 +127,9 @@ class GnocchiHelper(base.DataSourceBase):
|
|||||||
kwargs = {k: v for k, v in raw_kwargs.items() if k and v}
|
kwargs = {k: v for k, v in raw_kwargs.items() if k and v}
|
||||||
|
|
||||||
statistics = self.query_retry(
|
statistics = self.query_retry(
|
||||||
f=self.gnocchi.metric.get_measures, **kwargs)
|
f=self.gnocchi.metric.get_measures,
|
||||||
|
ignored_exc=gnc_exc.NotFound,
|
||||||
|
**kwargs)
|
||||||
|
|
||||||
return_value = None
|
return_value = None
|
||||||
if statistics:
|
if statistics:
|
||||||
@@ -120,6 +141,67 @@ class GnocchiHelper(base.DataSourceBase):
|
|||||||
# Airflow from hardware.ipmi.node.airflow is reported as
|
# Airflow from hardware.ipmi.node.airflow is reported as
|
||||||
# 1/10 th of actual CFM
|
# 1/10 th of actual CFM
|
||||||
return_value *= 10
|
return_value *= 10
|
||||||
|
if meter_name == "instance_cpu_usage":
|
||||||
|
# "rate:mean" can return negative values for migrated vms.
|
||||||
|
return_value = max(0, return_value)
|
||||||
|
|
||||||
|
# We're converting the cumulative cpu time (ns) to cpu usage
|
||||||
|
# percentage.
|
||||||
|
vcpus = resource.vcpus
|
||||||
|
if not vcpus:
|
||||||
|
LOG.warning("instance vcpu count not set, assuming 1")
|
||||||
|
vcpus = 1
|
||||||
|
return_value *= 100 / (granularity * 10e+8) / vcpus
|
||||||
|
|
||||||
|
return return_value
|
||||||
|
|
||||||
|
def statistic_series(self, resource=None, resource_type=None,
|
||||||
|
meter_name=None, start_time=None, end_time=None,
|
||||||
|
granularity=300):
|
||||||
|
|
||||||
|
meter = self._get_meter(meter_name)
|
||||||
|
|
||||||
|
resource_id = resource.uuid
|
||||||
|
if resource_type == 'compute_node':
|
||||||
|
resource_id = "%s_%s" % (resource.hostname, resource.hostname)
|
||||||
|
kwargs = dict(query={"=": {"original_resource_id": resource_id}},
|
||||||
|
limit=1)
|
||||||
|
resources = self.query_retry(
|
||||||
|
f=self.gnocchi.resource.search,
|
||||||
|
ignored_exc=gnc_exc.NotFound,
|
||||||
|
**kwargs)
|
||||||
|
|
||||||
|
if not resources:
|
||||||
|
LOG.warning("The {0} resource {1} could not be "
|
||||||
|
"found".format(self.NAME, resource_id))
|
||||||
|
return
|
||||||
|
|
||||||
|
resource_id = resources[0]['id']
|
||||||
|
|
||||||
|
raw_kwargs = dict(
|
||||||
|
metric=meter,
|
||||||
|
start=start_time,
|
||||||
|
stop=end_time,
|
||||||
|
resource_id=resource_id,
|
||||||
|
granularity=granularity,
|
||||||
|
)
|
||||||
|
|
||||||
|
kwargs = {k: v for k, v in raw_kwargs.items() if k and v}
|
||||||
|
|
||||||
|
statistics = self.query_retry(
|
||||||
|
f=self.gnocchi.metric.get_measures,
|
||||||
|
ignored_exc=gnc_exc.NotFound,
|
||||||
|
**kwargs)
|
||||||
|
|
||||||
|
return_value = None
|
||||||
|
if statistics:
|
||||||
|
# measure has structure [time, granularity, value]
|
||||||
|
if meter_name == 'host_airflow':
|
||||||
|
# Airflow from hardware.ipmi.node.airflow is reported as
|
||||||
|
# 1/10 th of actual CFM
|
||||||
|
return_value = {s[0]: s[2]*10 for s in statistics}
|
||||||
|
else:
|
||||||
|
return_value = {s[0]: s[2] for s in statistics}
|
||||||
|
|
||||||
return return_value
|
return return_value
|
||||||
|
|
||||||
|
|||||||
@@ -16,10 +16,13 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
|
from urllib import parse as urlparse
|
||||||
|
|
||||||
|
from http import HTTPStatus
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
import six.moves.urllib.parse as urlparse
|
|
||||||
|
|
||||||
|
from watcher._i18n import _
|
||||||
from watcher.common import clients
|
from watcher.common import clients
|
||||||
from watcher.common import exception
|
from watcher.common import exception
|
||||||
from watcher.decision_engine.datasources import base
|
from watcher.decision_engine.datasources import base
|
||||||
@@ -135,12 +138,13 @@ class GrafanaHelper(base.DataSourceBase):
|
|||||||
raise exception.DataSourceNotAvailable(self.NAME)
|
raise exception.DataSourceNotAvailable(self.NAME)
|
||||||
|
|
||||||
resp = requests.get(self._base_url + str(project_id) + '/query',
|
resp = requests.get(self._base_url + str(project_id) + '/query',
|
||||||
params=params, headers=self._headers)
|
params=params, headers=self._headers,
|
||||||
if resp.status_code == 200:
|
timeout=CONF.grafana_client.http_timeout)
|
||||||
|
if resp.status_code == HTTPStatus.OK:
|
||||||
return resp
|
return resp
|
||||||
elif resp.status_code == 400:
|
elif resp.status_code == HTTPStatus.BAD_REQUEST:
|
||||||
LOG.error("Query for metric is invalid")
|
LOG.error("Query for metric is invalid")
|
||||||
elif resp.status_code == 401:
|
elif resp.status_code == HTTPStatus.UNAUTHORIZED:
|
||||||
LOG.error("Authorization token is invalid")
|
LOG.error("Authorization token is invalid")
|
||||||
raise exception.DataSourceNotAvailable(self.NAME)
|
raise exception.DataSourceNotAvailable(self.NAME)
|
||||||
|
|
||||||
@@ -187,6 +191,12 @@ class GrafanaHelper(base.DataSourceBase):
|
|||||||
|
|
||||||
return result
|
return result
|
||||||
|
|
||||||
|
def statistic_series(self, resource=None, resource_type=None,
|
||||||
|
meter_name=None, start_time=None, end_time=None,
|
||||||
|
granularity=300):
|
||||||
|
raise NotImplementedError(
|
||||||
|
_('Grafana helper does not support statistic series method'))
|
||||||
|
|
||||||
def get_host_cpu_usage(self, resource, period=300,
|
def get_host_cpu_usage(self, resource, period=300,
|
||||||
aggregate="mean", granularity=None):
|
aggregate="mean", granularity=None):
|
||||||
return self.statistic_aggregation(
|
return self.statistic_aggregation(
|
||||||
|
|||||||
@@ -21,7 +21,6 @@ import datetime
|
|||||||
from monascaclient import exc
|
from monascaclient import exc
|
||||||
|
|
||||||
from watcher.common import clients
|
from watcher.common import clients
|
||||||
from watcher.common import exception
|
|
||||||
from watcher.decision_engine.datasources import base
|
from watcher.decision_engine.datasources import base
|
||||||
|
|
||||||
|
|
||||||
@@ -90,9 +89,7 @@ class MonascaHelper(base.DataSourceBase):
|
|||||||
stop_time = datetime.datetime.utcnow()
|
stop_time = datetime.datetime.utcnow()
|
||||||
start_time = stop_time - datetime.timedelta(seconds=(int(period)))
|
start_time = stop_time - datetime.timedelta(seconds=(int(period)))
|
||||||
|
|
||||||
meter = self.METRIC_MAP.get(meter_name)
|
meter = self._get_meter(meter_name)
|
||||||
if meter is None:
|
|
||||||
raise exception.MetricNotAvailable(metric=meter_name)
|
|
||||||
|
|
||||||
if aggregate == 'mean':
|
if aggregate == 'mean':
|
||||||
aggregate = 'avg'
|
aggregate = 'avg'
|
||||||
@@ -121,6 +118,34 @@ class MonascaHelper(base.DataSourceBase):
|
|||||||
|
|
||||||
return cpu_usage
|
return cpu_usage
|
||||||
|
|
||||||
|
def statistic_series(self, resource=None, resource_type=None,
|
||||||
|
meter_name=None, start_time=None, end_time=None,
|
||||||
|
granularity=300):
|
||||||
|
|
||||||
|
meter = self._get_meter(meter_name)
|
||||||
|
|
||||||
|
raw_kwargs = dict(
|
||||||
|
name=meter,
|
||||||
|
start_time=start_time.isoformat(),
|
||||||
|
end_time=end_time.isoformat(),
|
||||||
|
dimensions={'hostname': resource.uuid},
|
||||||
|
statistics='avg',
|
||||||
|
group_by='*',
|
||||||
|
)
|
||||||
|
|
||||||
|
kwargs = {k: v for k, v in raw_kwargs.items() if k and v}
|
||||||
|
|
||||||
|
statistics = self.query_retry(
|
||||||
|
f=self.monasca.metrics.list_statistics, **kwargs)
|
||||||
|
|
||||||
|
result = {}
|
||||||
|
for stat in statistics:
|
||||||
|
v_index = stat['columns'].index('avg')
|
||||||
|
t_index = stat['columns'].index('timestamp')
|
||||||
|
result.update({r[t_index]: r[v_index] for r in stat['statistics']})
|
||||||
|
|
||||||
|
return result
|
||||||
|
|
||||||
def get_host_cpu_usage(self, resource, period,
|
def get_host_cpu_usage(self, resource, period,
|
||||||
aggregate, granularity=None):
|
aggregate, granularity=None):
|
||||||
return self.statistic_aggregation(
|
return self.statistic_aggregation(
|
||||||
|
|||||||
@@ -15,13 +15,11 @@
|
|||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
import abc
|
import abc
|
||||||
import six
|
|
||||||
|
|
||||||
from watcher.common.loader import loadable
|
from watcher.common.loader import loadable
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class Goal(loadable.Loadable, metaclass=abc.ABCMeta):
|
||||||
class Goal(loadable.Loadable):
|
|
||||||
|
|
||||||
def __init__(self, config):
|
def __init__(self, config):
|
||||||
super(Goal, self).__init__(config)
|
super(Goal, self).__init__(config)
|
||||||
|
|||||||
@@ -27,11 +27,8 @@ import abc
|
|||||||
import jsonschema
|
import jsonschema
|
||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
|
|
||||||
import six
|
|
||||||
|
|
||||||
|
class EfficacySpecification(object, metaclass=abc.ABCMeta):
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
|
||||||
class EfficacySpecification(object):
|
|
||||||
|
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
self._indicators_specs = self.get_indicators_specifications()
|
self._indicators_specs = self.get_indicators_specifications()
|
||||||
|
|||||||
@@ -18,7 +18,6 @@ import abc
|
|||||||
import jsonschema
|
import jsonschema
|
||||||
from jsonschema import SchemaError
|
from jsonschema import SchemaError
|
||||||
from jsonschema import ValidationError
|
from jsonschema import ValidationError
|
||||||
import six
|
|
||||||
|
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
@@ -29,8 +28,7 @@ from watcher.common import exception
|
|||||||
LOG = log.getLogger(__name__)
|
LOG = log.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class IndicatorSpecification(object, metaclass=abc.ABCMeta):
|
||||||
class IndicatorSpecification(object):
|
|
||||||
|
|
||||||
def __init__(self, name=None, description=None, unit=None, required=True):
|
def __init__(self, name=None, description=None, unit=None, required=True):
|
||||||
self.name = name
|
self.name = name
|
||||||
|
|||||||
@@ -19,9 +19,6 @@
|
|||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
#
|
#
|
||||||
|
|
||||||
from __future__ import unicode_literals
|
|
||||||
|
|
||||||
|
|
||||||
from watcher.common.loader import default
|
from watcher.common.loader import default
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -25,11 +25,9 @@ See: :doc:`../architecture` for more details on this component.
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
import abc
|
import abc
|
||||||
import six
|
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class Model(object, metaclass=abc.ABCMeta):
|
||||||
class Model(object):
|
|
||||||
|
|
||||||
@abc.abstractmethod
|
@abc.abstractmethod
|
||||||
def to_string(self):
|
def to_string(self):
|
||||||
|
|||||||
@@ -110,7 +110,6 @@ import time
|
|||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
import six
|
|
||||||
|
|
||||||
from watcher.common import clients
|
from watcher.common import clients
|
||||||
from watcher.common.loader import loadable
|
from watcher.common.loader import loadable
|
||||||
@@ -120,8 +119,8 @@ LOG = log.getLogger(__name__)
|
|||||||
CONF = cfg.CONF
|
CONF = cfg.CONF
|
||||||
|
|
||||||
|
|
||||||
@six.add_metaclass(abc.ABCMeta)
|
class BaseClusterDataModelCollector(loadable.LoadableSingleton,
|
||||||
class BaseClusterDataModelCollector(loadable.LoadableSingleton):
|
metaclass=abc.ABCMeta):
|
||||||
|
|
||||||
STALE_MODEL = model_root.ModelRoot(stale=True)
|
STALE_MODEL = model_root.ModelRoot(stale=True)
|
||||||
|
|
||||||
|
|||||||
@@ -13,8 +13,6 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
import six
|
|
||||||
|
|
||||||
from oslo_log import log
|
from oslo_log import log
|
||||||
|
|
||||||
from watcher.common import cinder_helper
|
from watcher.common import cinder_helper
|
||||||
@@ -152,6 +150,9 @@ class CinderClusterDataModelCollector(base.BaseClusterDataModelCollector):
|
|||||||
if self._audit_scope_handler is None:
|
if self._audit_scope_handler is None:
|
||||||
LOG.debug("No audit, Don't Build storage data model")
|
LOG.debug("No audit, Don't Build storage data model")
|
||||||
return
|
return
|
||||||
|
if self._data_model_scope is None:
|
||||||
|
LOG.debug("No audit scope, Don't Build storage data model")
|
||||||
|
return
|
||||||
|
|
||||||
builder = CinderModelBuilder(self.osc)
|
builder = CinderModelBuilder(self.osc)
|
||||||
return builder.execute(self._data_model_scope)
|
return builder.execute(self._data_model_scope)
|
||||||
@@ -204,7 +205,7 @@ class CinderModelBuilder(base.BaseModelBuilder):
|
|||||||
"""Build a storage node from a Cinder storage node
|
"""Build a storage node from a Cinder storage node
|
||||||
|
|
||||||
:param node: A storage node
|
:param node: A storage node
|
||||||
:type node: :py:class:`~cinderclient.v2.services.Service`
|
:type node: :py:class:`~cinderclient.v3.services.Service`
|
||||||
"""
|
"""
|
||||||
# node.host is formatted as host@backendname since ocata,
|
# node.host is formatted as host@backendname since ocata,
|
||||||
# or may be only host as of ocata
|
# or may be only host as of ocata
|
||||||
@@ -232,7 +233,7 @@ class CinderModelBuilder(base.BaseModelBuilder):
|
|||||||
"""Build a storage pool from a Cinder storage pool
|
"""Build a storage pool from a Cinder storage pool
|
||||||
|
|
||||||
:param pool: A storage pool
|
:param pool: A storage pool
|
||||||
:type pool: :py:class:`~cinderclient.v2.pools.Pool`
|
:type pool: :py:class:`~cinderclient.v3.pools.Pool`
|
||||||
:raises: exception.InvalidPoolAttributeValue
|
:raises: exception.InvalidPoolAttributeValue
|
||||||
"""
|
"""
|
||||||
# build up the storage pool.
|
# build up the storage pool.
|
||||||
@@ -286,7 +287,7 @@ class CinderModelBuilder(base.BaseModelBuilder):
|
|||||||
:param instance: Cinder Volume object.
|
:param instance: Cinder Volume object.
|
||||||
:return: A volume node for the graph.
|
:return: A volume node for the graph.
|
||||||
"""
|
"""
|
||||||
attachments = [{k: v for k, v in six.iteritems(d) if k in (
|
attachments = [{k: v for k, v in iter(d.items()) if k in (
|
||||||
'server_id', 'attachment_id')} for d in volume.attachments]
|
'server_id', 'attachment_id')} for d in volume.attachments]
|
||||||
|
|
||||||
volume_attributes = {
|
volume_attributes = {
|
||||||
|
|||||||
@@ -63,6 +63,9 @@ class BaremetalClusterDataModelCollector(base.BaseClusterDataModelCollector):
|
|||||||
if self._audit_scope_handler is None:
|
if self._audit_scope_handler is None:
|
||||||
LOG.debug("No audit, Don't Build Baremetal data model")
|
LOG.debug("No audit, Don't Build Baremetal data model")
|
||||||
return
|
return
|
||||||
|
if self._data_model_scope is None:
|
||||||
|
LOG.debug("No audit scope, Don't Build Baremetal data model")
|
||||||
|
return
|
||||||
|
|
||||||
builder = BareMetalModelBuilder(self.osc)
|
builder = BareMetalModelBuilder(self.osc)
|
||||||
return builder.execute(self._data_model_scope)
|
return builder.execute(self._data_model_scope)
|
||||||
@@ -78,6 +81,7 @@ class BareMetalModelBuilder(base.BaseModelBuilder):
|
|||||||
def __init__(self, osc):
|
def __init__(self, osc):
|
||||||
self.osc = osc
|
self.osc = osc
|
||||||
self.model = model_root.BaremetalModelRoot()
|
self.model = model_root.BaremetalModelRoot()
|
||||||
|
# TODO(lpetrut): add MAAS support
|
||||||
self.ironic_helper = ironic_helper.IronicHelper(osc=self.osc)
|
self.ironic_helper = ironic_helper.IronicHelper(osc=self.osc)
|
||||||
|
|
||||||
def add_ironic_node(self, node):
|
def add_ironic_node(self, node):
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user