Compare commits
72 Commits
4.0.0.0rc2
...
wallaby-em
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
d60e2a96e4 | ||
|
|
adf4725f8c | ||
|
|
537823b216 | ||
|
|
281455a08a | ||
|
|
863815153e | ||
|
|
76270c8383 | ||
|
|
58de9c405a | ||
|
|
8f0126f1fe | ||
|
|
ec21898978 | ||
|
|
e61f9b5e88 | ||
|
|
e91efbde01 | ||
|
|
63b6997c83 | ||
|
|
262edc8cc9 | ||
|
|
204b276693 | ||
|
|
f8a2877f24 | ||
|
|
af02bebca9 | ||
|
|
3aaa20908d | ||
|
|
5097665be3 | ||
|
|
09f6e3bde5 | ||
|
|
f488636fb8 | ||
|
|
11cb88c2cd | ||
|
|
16a0486655 | ||
|
|
2454d4d199 | ||
|
|
45dca00dee | ||
|
|
09b2383685 | ||
|
|
f8797a7f70 | ||
|
|
da283b49b8 | ||
|
|
e21e5f609e | ||
|
|
583c946061 | ||
|
|
cca0d9f7d7 | ||
|
|
25a0b184a1 | ||
|
|
ed59145354 | ||
|
|
19adfda3b9 | ||
|
|
fa56bc715e | ||
|
|
350ce66d3c | ||
|
|
1667046f58 | ||
|
|
3f7a508a2e | ||
|
|
f7f5659bca | ||
|
|
57f55190ff | ||
|
|
237550ad57 | ||
|
|
cad67702d6 | ||
|
|
ae678dfaaa | ||
|
|
5ad3960286 | ||
|
|
dbd86be363 | ||
|
|
9f0138e1cf | ||
|
|
097ac06f0b | ||
|
|
0869b1c75c | ||
|
|
527578a147 | ||
|
|
b0c411b22a | ||
|
|
4a1915bec4 | ||
|
|
751027858b | ||
|
|
12bd9c0590 | ||
|
|
1ff940598f | ||
|
|
9d495618d2 | ||
|
|
c6d2690aa3 | ||
|
|
623e44ecf9 | ||
|
|
5c34b6bc47 | ||
|
|
8a36ad5f87 | ||
|
|
6ff95efaf6 | ||
|
|
ba2f1804b0 | ||
|
|
44061326e9 | ||
|
|
0b4c4f1de6 | ||
|
|
9652571437 | ||
|
|
f0f15f89c6 | ||
|
|
075e374b3d | ||
|
|
eaa0dfea4b | ||
|
|
b7956de761 | ||
|
|
a30dbdd724 | ||
|
|
60a829e982 | ||
|
|
74cfa0fc8c | ||
|
|
5071c8f8fa | ||
|
|
0ef0f165cb |
@@ -2,3 +2,4 @@
|
||||
host=review.opendev.org
|
||||
port=29418
|
||||
project=openstack/watcher.git
|
||||
defaultbranch=stable/wallaby
|
||||
|
||||
30
.zuul.yaml
30
.zuul.yaml
@@ -2,8 +2,7 @@
|
||||
templates:
|
||||
- check-requirements
|
||||
- openstack-cover-jobs
|
||||
- openstack-lower-constraints-jobs
|
||||
- openstack-python3-ussuri-jobs
|
||||
- openstack-python3-wallaby-jobs
|
||||
- publish-openstack-docs-pti
|
||||
- release-notes-jobs-python3
|
||||
check:
|
||||
@@ -102,7 +101,7 @@
|
||||
- job:
|
||||
name: watcher-tempest-multinode
|
||||
parent: watcher-tempest-functional
|
||||
nodeset: openstack-two-node-bionic
|
||||
nodeset: openstack-two-node-focal
|
||||
roles:
|
||||
- zuul: openstack/tempest
|
||||
group-vars:
|
||||
@@ -160,7 +159,6 @@
|
||||
timeout: 7200
|
||||
required-projects: &base_required_projects
|
||||
- openstack/ceilometer
|
||||
- openstack/devstack-gate
|
||||
- openstack/python-openstackclient
|
||||
- openstack/python-watcherclient
|
||||
- openstack/watcher
|
||||
@@ -179,14 +177,10 @@
|
||||
s-container: false
|
||||
s-object: false
|
||||
s-proxy: false
|
||||
devstack_localrc:
|
||||
TEMPEST_PLUGINS: /opt/stack/watcher-tempest-plugin
|
||||
USE_PYTHON3: true
|
||||
tempest_plugins:
|
||||
- watcher-tempest-plugin
|
||||
tempest_test_regex: watcher_tempest_plugin.tests.api
|
||||
tox_envlist: all
|
||||
tox_environment:
|
||||
# Do we really need to set this? It's cargo culted
|
||||
PYTHONUNBUFFERED: 'true'
|
||||
zuul_copy_output:
|
||||
/etc/hosts: logs
|
||||
|
||||
@@ -200,10 +194,12 @@
|
||||
|
||||
- job:
|
||||
name: watcher-grenade
|
||||
parent: legacy-dsvm-base
|
||||
timeout: 10800
|
||||
run: playbooks/legacy/grenade-devstack-watcher/run.yaml
|
||||
post-run: playbooks/legacy/grenade-devstack-watcher/post.yaml
|
||||
parent: grenade
|
||||
required-projects:
|
||||
- openstack/watcher
|
||||
- openstack/python-watcherclient
|
||||
- openstack/watcher-tempest-plugin
|
||||
vars: *base_vars
|
||||
irrelevant-files:
|
||||
- ^(test-|)requirements.txt$
|
||||
- ^.*\.rst$
|
||||
@@ -215,12 +211,6 @@
|
||||
- ^setup.cfg$
|
||||
- ^tools/.*$
|
||||
- ^tox.ini$
|
||||
required-projects:
|
||||
- openstack/grenade
|
||||
- openstack/devstack-gate
|
||||
- openstack/watcher
|
||||
- openstack/python-watcherclient
|
||||
- openstack/watcher-tempest-plugin
|
||||
|
||||
- job:
|
||||
# This job is used in python-watcherclient repo
|
||||
|
||||
@@ -22,9 +22,6 @@
|
||||
# All configuration values have a default; values that are commented out
|
||||
# serve to show the default.
|
||||
|
||||
from watcher import version as watcher_version
|
||||
|
||||
|
||||
extensions = [
|
||||
'openstackdocstheme',
|
||||
'os_api_ref',
|
||||
@@ -46,21 +43,13 @@ project = u'Infrastructure Optimization API Reference'
|
||||
copyright = u'2010-present, OpenStack Foundation'
|
||||
|
||||
# openstackdocstheme options
|
||||
repository_name = 'openstack/watcher'
|
||||
bug_project = 'watcher'
|
||||
bug_tag = ''
|
||||
|
||||
# The version info for the project you're documenting, acts as replacement for
|
||||
# |version| and |release|, also used in various other places throughout the
|
||||
# built documents.
|
||||
#
|
||||
# The full version, including alpha/beta/rc tags.
|
||||
release = watcher_version.version_info.release_string()
|
||||
# The short X.Y version.
|
||||
version = watcher_version.version_string
|
||||
openstackdocs_repo_name = 'openstack/watcher'
|
||||
openstackdocs_auto_name = False
|
||||
openstackdocs_bug_project = 'watcher'
|
||||
openstackdocs_bug_tag = ''
|
||||
|
||||
# The name of the Pygments (syntax highlighting) style to use.
|
||||
pygments_style = 'sphinx'
|
||||
pygments_style = 'native'
|
||||
|
||||
# -- Options for HTML output --------------------------------------------------
|
||||
|
||||
@@ -75,10 +64,6 @@ html_theme_options = {
|
||||
"sidebar_mode": "toc",
|
||||
}
|
||||
|
||||
# If not '', a 'Last updated on:' timestamp is inserted at every page bottom,
|
||||
# using the given strftime format.
|
||||
html_last_updated_fmt = '%Y-%m-%d %H:%M'
|
||||
|
||||
# -- Options for LaTeX output -------------------------------------------------
|
||||
|
||||
# Grouping the document tree into LaTeX files. List of tuples
|
||||
|
||||
@@ -298,7 +298,7 @@ function start_watcher_api {
|
||||
service_protocol="http"
|
||||
fi
|
||||
if [[ "$WATCHER_USE_WSGI_MODE" == "uwsgi" ]]; then
|
||||
run_process "watcher-api" "$WATCHER_BIN_DIR/uwsgi --procname-prefix watcher-api --ini $WATCHER_UWSGI_CONF"
|
||||
run_process "watcher-api" "$(which uwsgi) --procname-prefix watcher-api --ini $WATCHER_UWSGI_CONF"
|
||||
watcher_url=$service_protocol://$SERVICE_HOST/infra-optim
|
||||
else
|
||||
watcher_url=$service_protocol://$SERVICE_HOST:$service_port
|
||||
|
||||
@@ -13,8 +13,6 @@
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
from __future__ import unicode_literals
|
||||
|
||||
import importlib
|
||||
import inspect
|
||||
|
||||
|
||||
@@ -1,10 +1,10 @@
|
||||
# The order of packages is significant, because pip processes them in the order
|
||||
# of appearance. Changing the order has an impact on the overall integration
|
||||
# process, which may cause wedges in the gate later.
|
||||
openstackdocstheme>=1.20.0 # Apache-2.0
|
||||
sphinx>=1.8.0,!=2.1.0,!=3.0.0 # BSD
|
||||
openstackdocstheme>=2.2.1 # Apache-2.0
|
||||
sphinx>=2.0.0,!=2.1.0 # BSD
|
||||
sphinxcontrib-pecanwsme>=0.8.0 # Apache-2.0
|
||||
sphinxcontrib-svg2pdfconverter>=0.1.0 # BSD
|
||||
reno>=2.7.0 # Apache-2.0
|
||||
reno>=3.1.0 # Apache-2.0
|
||||
sphinxcontrib-apidoc>=0.2.0 # BSD
|
||||
os-api-ref>=1.4.0 # Apache-2.0
|
||||
|
||||
@@ -17,6 +17,14 @@
|
||||
Policies
|
||||
========
|
||||
|
||||
.. warning::
|
||||
|
||||
JSON formatted policy file is deprecated since Watcher 6.0.0 (Wallaby).
|
||||
This `oslopolicy-convert-json-to-yaml`__ tool will migrate your existing
|
||||
JSON-formatted policy file to YAML in a backward-compatible way.
|
||||
|
||||
.. __: https://docs.openstack.org/oslo.policy/latest/cli/oslopolicy-convert-json-to-yaml.html
|
||||
|
||||
Watcher's public API calls may be restricted to certain sets of users using a
|
||||
policy configuration file. This document explains exactly how policies are
|
||||
configured and what they apply to.
|
||||
|
||||
@@ -14,7 +14,6 @@
|
||||
import os
|
||||
import sys
|
||||
|
||||
from watcher import version as watcher_version
|
||||
from watcher import objects
|
||||
|
||||
objects.register_all()
|
||||
@@ -60,16 +59,6 @@ master_doc = 'index'
|
||||
project = u'Watcher'
|
||||
copyright = u'OpenStack Foundation'
|
||||
|
||||
# The version info for the project you're documenting, acts as replacement for
|
||||
# |version| and |release|, also used in various other places throughout the
|
||||
# built documents.
|
||||
#
|
||||
# The short X.Y version.
|
||||
# The full version, including alpha/beta/rc tags.
|
||||
release = watcher_version.version_info.release_string()
|
||||
# The short X.Y version.
|
||||
version = watcher_version.version_string
|
||||
|
||||
# A list of ignored prefixes for module index sorting.
|
||||
modindex_common_prefix = ['watcher.']
|
||||
|
||||
@@ -94,7 +83,7 @@ add_module_names = True
|
||||
suppress_warnings = ['app.add_directive']
|
||||
|
||||
# The name of the Pygments (syntax highlighting) style to use.
|
||||
pygments_style = 'sphinx'
|
||||
pygments_style = 'native'
|
||||
|
||||
# -- Options for man page output --------------------------------------------
|
||||
|
||||
@@ -125,12 +114,13 @@ html_theme = 'openstackdocs'
|
||||
# Output file base name for HTML help builder.
|
||||
htmlhelp_basename = '%sdoc' % project
|
||||
|
||||
html_last_updated_fmt = '%Y-%m-%d %H:%M'
|
||||
|
||||
#openstackdocstheme options
|
||||
repository_name = 'openstack/watcher'
|
||||
bug_project = 'watcher'
|
||||
bug_tag = ''
|
||||
openstackdocs_repo_name = 'openstack/watcher'
|
||||
openstackdocs_pdf_link = True
|
||||
openstackdocs_auto_name = False
|
||||
openstackdocs_bug_project = 'watcher'
|
||||
openstackdocs_bug_tag = ''
|
||||
|
||||
# Grouping the document tree into LaTeX files. List of tuples
|
||||
# (source start file, target name, title, author, documentclass
|
||||
@@ -138,7 +128,7 @@ bug_tag = ''
|
||||
latex_documents = [
|
||||
('index',
|
||||
'doc-watcher.tex',
|
||||
u'%s Documentation' % project,
|
||||
u'Watcher Documentation',
|
||||
u'OpenStack Foundation', 'manual'),
|
||||
]
|
||||
|
||||
|
||||
@@ -56,9 +56,6 @@ Here is an example showing how you can write a plugin called ``NewStrategy``:
|
||||
# filepath: thirdparty/new.py
|
||||
# import path: thirdparty.new
|
||||
import abc
|
||||
|
||||
import six
|
||||
|
||||
from watcher._i18n import _
|
||||
from watcher.decision_engine.strategy.strategies import base
|
||||
|
||||
|
||||
@@ -1,161 +0,0 @@
|
||||
alabaster==0.7.10
|
||||
alembic==0.9.8
|
||||
amqp==2.2.2
|
||||
appdirs==1.4.3
|
||||
APScheduler==3.5.1
|
||||
asn1crypto==0.24.0
|
||||
automaton==1.14.0
|
||||
Babel==2.5.3
|
||||
beautifulsoup4==4.6.0
|
||||
cachetools==2.0.1
|
||||
certifi==2018.1.18
|
||||
cffi==1.11.5
|
||||
chardet==3.0.4
|
||||
cliff==2.11.0
|
||||
cmd2==0.8.1
|
||||
contextlib2==0.5.5
|
||||
coverage==4.5.1
|
||||
croniter==0.3.20
|
||||
cryptography==2.1.4
|
||||
debtcollector==1.19.0
|
||||
decorator==4.2.1
|
||||
deprecation==2.0
|
||||
doc8==0.8.0
|
||||
docutils==0.14
|
||||
dogpile.cache==0.6.5
|
||||
dulwich==0.19.0
|
||||
enum34==1.1.6
|
||||
enum-compat==0.0.2
|
||||
eventlet==0.20.0
|
||||
extras==1.0.0
|
||||
fasteners==0.14.1
|
||||
fixtures==3.0.0
|
||||
freezegun==0.3.10
|
||||
future==0.16.0
|
||||
futurist==1.8.0
|
||||
gitdb2==2.0.3
|
||||
GitPython==2.1.8
|
||||
gnocchiclient==7.0.1
|
||||
greenlet==0.4.13
|
||||
idna==2.6
|
||||
imagesize==1.0.0
|
||||
iso8601==0.1.12
|
||||
Jinja2==2.10
|
||||
jmespath==0.9.3
|
||||
jsonpatch==1.21
|
||||
jsonpointer==2.0
|
||||
jsonschema==2.6.0
|
||||
keystoneauth1==3.4.0
|
||||
keystonemiddleware==4.21.0
|
||||
kombu==4.1.0
|
||||
linecache2==1.0.0
|
||||
logutils==0.3.5
|
||||
lxml==4.1.1
|
||||
Mako==1.0.7
|
||||
MarkupSafe==1.0
|
||||
mccabe==0.2.1
|
||||
microversion_parse==0.2.1
|
||||
mock==2.0.0
|
||||
monotonic==1.4
|
||||
mox3==0.25.0
|
||||
msgpack==0.5.6
|
||||
munch==2.2.0
|
||||
netaddr==0.7.19
|
||||
netifaces==0.10.6
|
||||
networkx==2.2
|
||||
openstackdocstheme==1.20.0
|
||||
openstacksdk==0.12.0
|
||||
os-api-ref===1.4.0
|
||||
os-client-config==1.29.0
|
||||
os-service-types==1.2.0
|
||||
os-testr==1.0.0
|
||||
osc-lib==1.10.0
|
||||
os-resource-classes==0.4.0
|
||||
oslo.cache==1.29.0
|
||||
oslo.concurrency==3.26.0
|
||||
oslo.config==5.2.0
|
||||
oslo.context==2.21.0
|
||||
oslo.db==4.35.0
|
||||
oslo.i18n==3.20.0
|
||||
oslo.log==3.37.0
|
||||
oslo.messaging==8.1.2
|
||||
oslo.middleware==3.35.0
|
||||
oslo.policy==1.34.0
|
||||
oslo.reports==1.27.0
|
||||
oslo.serialization==2.25.0
|
||||
oslo.service==1.30.0
|
||||
oslo.upgradecheck==0.1.0
|
||||
oslo.utils==3.36.0
|
||||
oslo.versionedobjects==1.32.0
|
||||
oslotest==3.3.0
|
||||
packaging==17.1
|
||||
Paste==2.0.3
|
||||
PasteDeploy==1.5.2
|
||||
pbr==3.1.1
|
||||
pecan==1.3.2
|
||||
pika==0.10.0
|
||||
pika-pool==0.1.3
|
||||
prettytable==0.7.2
|
||||
psutil==5.4.3
|
||||
pycadf==2.7.0
|
||||
pycparser==2.18
|
||||
Pygments==2.2.0
|
||||
pyinotify==0.9.6
|
||||
pyOpenSSL==17.5.0
|
||||
pyparsing==2.2.0
|
||||
pyperclip==1.6.0
|
||||
python-ceilometerclient==2.9.0
|
||||
python-cinderclient==3.5.0
|
||||
python-dateutil==2.7.0
|
||||
python-editor==1.0.3
|
||||
python-glanceclient==2.9.1
|
||||
python-ironicclient==2.5.0
|
||||
python-keystoneclient==3.15.0
|
||||
python-mimeparse==1.6.0
|
||||
python-monascaclient==1.12.0
|
||||
python-neutronclient==6.7.0
|
||||
python-novaclient==14.1.0
|
||||
python-openstackclient==3.14.0
|
||||
python-subunit==1.2.0
|
||||
pytz==2018.3
|
||||
PyYAML==3.12
|
||||
reno==2.7.0
|
||||
repoze.lru==0.7
|
||||
requests==2.18.4
|
||||
requestsexceptions==1.4.0
|
||||
restructuredtext-lint==1.1.3
|
||||
rfc3986==1.1.0
|
||||
Routes==2.4.1
|
||||
simplegeneric==0.8.1
|
||||
simplejson==3.13.2
|
||||
six==1.11.0
|
||||
smmap2==2.0.3
|
||||
snowballstemmer==1.2.1
|
||||
Sphinx==1.6.5
|
||||
sphinxcontrib-httpdomain==1.6.1
|
||||
sphinxcontrib-pecanwsme==0.8.0
|
||||
sphinxcontrib-websupport==1.0.1
|
||||
SQLAlchemy==1.2.5
|
||||
sqlalchemy-migrate==0.11.0
|
||||
sqlparse==0.2.4
|
||||
statsd==3.2.2
|
||||
stestr==2.0.0
|
||||
stevedore==1.28.0
|
||||
taskflow==3.7.1
|
||||
Tempita==0.5.2
|
||||
tenacity==4.9.0
|
||||
testresources==2.0.1
|
||||
testscenarios==0.5.0
|
||||
testtools==2.3.0
|
||||
traceback2==1.4.0
|
||||
tzlocal==1.5.1
|
||||
ujson==1.35
|
||||
unittest2==1.1.0
|
||||
urllib3==1.22
|
||||
vine==1.1.4
|
||||
waitress==1.1.0
|
||||
warlock==1.3.0
|
||||
WebOb==1.8.5
|
||||
WebTest==2.0.29
|
||||
wrapt==1.10.11
|
||||
WSME==0.9.2
|
||||
@@ -1,15 +0,0 @@
|
||||
- hosts: primary
|
||||
tasks:
|
||||
|
||||
- name: Copy files from {{ ansible_user_dir }}/workspace/ on node
|
||||
synchronize:
|
||||
src: '{{ ansible_user_dir }}/workspace/'
|
||||
dest: '{{ zuul.executor.log_root }}'
|
||||
mode: pull
|
||||
copy_links: true
|
||||
verify_host: true
|
||||
rsync_opts:
|
||||
- --include=/logs/**
|
||||
- --include=*/
|
||||
- --exclude=*
|
||||
- --prune-empty-dirs
|
||||
@@ -1,60 +0,0 @@
|
||||
- hosts: all
|
||||
name: legacy-grenade-dsvm-watcher
|
||||
tasks:
|
||||
|
||||
- name: Ensure legacy workspace directory
|
||||
file:
|
||||
path: '{{ ansible_user_dir }}/workspace'
|
||||
state: directory
|
||||
|
||||
- shell:
|
||||
cmd: |
|
||||
set -e
|
||||
set -x
|
||||
cat > clonemap.yaml << EOF
|
||||
clonemap:
|
||||
- name: openstack/devstack-gate
|
||||
dest: devstack-gate
|
||||
EOF
|
||||
/usr/zuul-env/bin/zuul-cloner -m clonemap.yaml --cache-dir /opt/git \
|
||||
https://opendev.org \
|
||||
openstack/devstack-gate
|
||||
executable: /bin/bash
|
||||
chdir: '{{ ansible_user_dir }}/workspace'
|
||||
environment: '{{ zuul | zuul_legacy_vars }}'
|
||||
|
||||
- shell:
|
||||
cmd: |
|
||||
set -e
|
||||
set -x
|
||||
export PYTHONUNBUFFERED=true
|
||||
|
||||
export PROJECTS="openstack/grenade $PROJECTS"
|
||||
export PROJECTS="openstack/watcher $PROJECTS"
|
||||
export PROJECTS="openstack/watcher-tempest-plugin $PROJECTS"
|
||||
export PROJECTS="openstack/python-watcherclient $PROJECTS"
|
||||
export DEVSTACK_PROJECT_FROM_GIT="python-watcherclient $DEVSTACK_PROJECT_FROM_GIT"
|
||||
|
||||
export GRENADE_PLUGINRC="enable_grenade_plugin watcher https://opendev.org/openstack/watcher"
|
||||
export DEVSTACK_LOCAL_CONFIG+=$'\n'"export TEMPEST_PLUGINS='/opt/stack/new/watcher-tempest-plugin'"
|
||||
|
||||
export DEVSTACK_GATE_TEMPEST_NOTESTS=1
|
||||
export DEVSTACK_GATE_GRENADE=pullup
|
||||
export DEVSTACK_GATE_USE_PYTHON3=True
|
||||
export BRANCH_OVERRIDE=default
|
||||
if [ "$BRANCH_OVERRIDE" != "default" ] ; then
|
||||
export OVERRIDE_ZUUL_BRANCH=$BRANCH_OVERRIDE
|
||||
fi
|
||||
# Add configuration values for enabling security features in local.conf
|
||||
function pre_test_hook {
|
||||
if [ -f /opt/stack/old/watcher-tempest-plugin/tools/pre_test_hook.sh ] ; then
|
||||
. /opt/stack/old/watcher-tempest-plugin/tools/pre_test_hook.sh
|
||||
fi
|
||||
}
|
||||
export -f pre_test_hook
|
||||
|
||||
cp devstack-gate/devstack-vm-gate-wrap.sh ./safe-devstack-vm-gate-wrap.sh
|
||||
./safe-devstack-vm-gate-wrap.sh
|
||||
executable: /bin/bash
|
||||
chdir: '{{ ansible_user_dir }}/workspace'
|
||||
environment: '{{ zuul | zuul_legacy_vars }}'
|
||||
@@ -0,0 +1,20 @@
|
||||
---
|
||||
upgrade:
|
||||
- |
|
||||
The default value of ``[oslo_policy] policy_file`` config option has
|
||||
been changed from ``policy.json`` to ``policy.yaml``.
|
||||
Operators who are utilizing customized or previously generated
|
||||
static policy JSON files (which are not needed by default), should
|
||||
generate new policy files or convert them in YAML format. Use the
|
||||
`oslopolicy-convert-json-to-yaml
|
||||
<https://docs.openstack.org/oslo.policy/latest/cli/oslopolicy-convert-json-to-yaml.html>`_
|
||||
tool to convert a JSON to YAML formatted policy file in
|
||||
backward compatible way.
|
||||
deprecations:
|
||||
- |
|
||||
Use of JSON policy files was deprecated by the ``oslo.policy`` library
|
||||
during the Victoria development cycle. As a result, this deprecation is
|
||||
being noted in the Wallaby cycle with an anticipated future removal of support
|
||||
by ``oslo.policy``. As such operators will need to convert to YAML policy
|
||||
files. Please see the upgrade notes for details on migration of any
|
||||
custom policy files.
|
||||
@@ -53,7 +53,6 @@ source_suffix = '.rst'
|
||||
master_doc = 'index'
|
||||
|
||||
# General information about the project.
|
||||
project = u'watcher'
|
||||
copyright = u'2016, Watcher developers'
|
||||
|
||||
# Release notes are version independent
|
||||
@@ -91,11 +90,15 @@ exclude_patterns = ['_build']
|
||||
#show_authors = False
|
||||
|
||||
# The name of the Pygments (syntax highlighting) style to use.
|
||||
pygments_style = 'sphinx'
|
||||
pygments_style = 'native'
|
||||
|
||||
# A list of ignored prefixes for module index sorting.
|
||||
#modindex_common_prefix = []
|
||||
|
||||
# openstackdocstheme options
|
||||
openstackdocs_repo_name = 'openstack/watcher'
|
||||
openstackdocs_bug_project = 'watcher'
|
||||
openstackdocs_bug_tag = ''
|
||||
|
||||
# -- Options for HTML output --------------------------------------------------
|
||||
|
||||
|
||||
@@ -21,6 +21,8 @@ Contents:
|
||||
:maxdepth: 1
|
||||
|
||||
unreleased
|
||||
victoria
|
||||
ussuri
|
||||
train
|
||||
stein
|
||||
rocky
|
||||
|
||||
@@ -1,14 +1,15 @@
|
||||
# Andi Chandler <andi@gowling.com>, 2017. #zanata
|
||||
# Andi Chandler <andi@gowling.com>, 2018. #zanata
|
||||
# Andi Chandler <andi@gowling.com>, 2020. #zanata
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: python-watcher\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2018-11-08 01:22+0000\n"
|
||||
"POT-Creation-Date: 2020-10-27 04:13+0000\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=UTF-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"PO-Revision-Date: 2018-11-07 06:15+0000\n"
|
||||
"PO-Revision-Date: 2020-10-28 11:13+0000\n"
|
||||
"Last-Translator: Andi Chandler <andi@gowling.com>\n"
|
||||
"Language-Team: English (United Kingdom)\n"
|
||||
"Language: en_GB\n"
|
||||
@@ -54,6 +55,61 @@ msgstr "1.7.0"
|
||||
msgid "1.9.0"
|
||||
msgstr "1.9.0"
|
||||
|
||||
msgid "2.0.0"
|
||||
msgstr "2.0.0"
|
||||
|
||||
msgid "3.0.0"
|
||||
msgstr "3.0.0"
|
||||
|
||||
msgid "4.0.0"
|
||||
msgstr "4.0.0"
|
||||
|
||||
msgid "A ``watcher-status upgrade check`` has been added for this."
|
||||
msgstr "A ``watcher-status upgrade check`` has been added for this."
|
||||
|
||||
msgid ""
|
||||
"A new threadpool for the decision engine that contributors can use to "
|
||||
"improve the performance of many operations, primarily I/O bound onces. The "
|
||||
"amount of workers used by the decision engine threadpool can be configured "
|
||||
"to scale according to the available infrastructure using the "
|
||||
"`watcher_decision_engine.max_general_workers` config option. Documentation "
|
||||
"for contributors to effectively use this threadpool is available online: "
|
||||
"https://docs.openstack.org/watcher/latest/contributor/concurrency.html"
|
||||
msgstr ""
|
||||
"A new threadpool for the decision engine that contributors can use to "
|
||||
"improve the performance of many operations, primarily I/O bound onces. The "
|
||||
"amount of workers used by the decision engine threadpool can be configured "
|
||||
"to scale according to the available infrastructure using the "
|
||||
"`watcher_decision_engine.max_general_workers` config option. Documentation "
|
||||
"for contributors to effectively use this threadpool is available online: "
|
||||
"https://docs.openstack.org/watcher/latest/contributor/concurrency.html"
|
||||
|
||||
msgid ""
|
||||
"API calls while building the Compute data model will be retried upon "
|
||||
"failure. The amount of failures allowed before giving up and the time before "
|
||||
"reattempting are configurable. The `api_call_retries` and "
|
||||
"`api_query_timeout` parameters in the `[collector]` group can be used to "
|
||||
"adjust these paremeters. 10 retries with a 1 second time in between "
|
||||
"reattempts is the default."
|
||||
msgstr ""
|
||||
"API calls while building the Compute data model will be retried upon "
|
||||
"failure. The amount of failures allowed before giving up and the time before "
|
||||
"reattempting are configurable. The `api_call_retries` and "
|
||||
"`api_query_timeout` parameters in the `[collector]` group can be used to "
|
||||
"adjust these parameters. 10 retries with a 1 second time in between "
|
||||
"reattempts is the default."
|
||||
|
||||
msgid ""
|
||||
"Add a new webhook API and a new audit type EVENT, the microversion is 1.4. "
|
||||
"Now Watcher user can create audit with EVENT type and the audit will be "
|
||||
"triggered by webhook API. The user guide is available online: https://docs."
|
||||
"openstack.org/watcher/latest/user/event_type_audit.html"
|
||||
msgstr ""
|
||||
"Add a new webhook API and a new audit type EVENT, the microversion is 1.4. "
|
||||
"Now Watcher user can create audit with EVENT type and the audit will be "
|
||||
"triggered by webhook API. The user guide is available online: https://docs."
|
||||
"openstack.org/watcher/latest/user/event_type_audit.html"
|
||||
|
||||
msgid "Add a service supervisor to watch Watcher deamons."
|
||||
msgstr "Add a service supervisor to watch Watcher daemons."
|
||||
|
||||
@@ -67,6 +123,24 @@ msgstr ""
|
||||
"Add description property for dynamic action. Admin can see detail "
|
||||
"information of any specify action."
|
||||
|
||||
msgid ""
|
||||
"Add force field to Audit. User can set --force to enable the new option when "
|
||||
"launching audit. If force is True, audit will be executed despite of ongoing "
|
||||
"actionplan. The new audit may create a wrong actionplan if they use the same "
|
||||
"data model."
|
||||
msgstr ""
|
||||
"Add force field to Audit. User can set --force to enable the new option when "
|
||||
"launching audit. If force is True, audit will be executed despite of ongoing "
|
||||
"actionplan. The new audit may create a wrong actionplan if they use the same "
|
||||
"data model."
|
||||
|
||||
msgid ""
|
||||
"Add keystone_client Group for user to configure 'interface' and "
|
||||
"'region_name' by watcher.conf. The default value of 'interface' is 'admin'."
|
||||
msgstr ""
|
||||
"Add keystone_client Group for user to configure 'interface' and "
|
||||
"'region_name' by watcher.conf. The default value of 'interface' is 'admin'."
|
||||
|
||||
msgid "Add notifications related to Action object."
|
||||
msgstr "Add notifications related to Action object."
|
||||
|
||||
@@ -79,6 +153,25 @@ msgstr "Add notifications related to Audit object."
|
||||
msgid "Add notifications related to Service object."
|
||||
msgstr "Add notifications related to Service object."
|
||||
|
||||
msgid ""
|
||||
"Add show data model api for Watcher. New in version 1.3. User can use "
|
||||
"'openstack optimize datamodel list' command to view the current data model "
|
||||
"information in memory. User can also add '--audit <Audit_UUID>' to view "
|
||||
"specific data model in memory filted by the scope in audit. User can also "
|
||||
"add '--detail' to view detailed information about current data model. User "
|
||||
"can also add '--type <type>' to specify the type of data model. Default type "
|
||||
"is 'compute'. In the future, type 'storage' and 'baremetal' will be "
|
||||
"supported."
|
||||
msgstr ""
|
||||
"Add show data model API for Watcher. New in version 1.3. User can use "
|
||||
"'openstack optimize datamodel list' command to view the current data model "
|
||||
"information in memory. User can also add '--audit <Audit_UUID>' to view "
|
||||
"specific data model in memory filtered by the scope in audit. User can also "
|
||||
"add '--detail' to view detailed information about current data model. User "
|
||||
"can also add '--type <type>' to specify the type of data model. Default type "
|
||||
"is 'compute'. In the future, type 'storage' and 'baremetal' will be "
|
||||
"supported."
|
||||
|
||||
msgid ""
|
||||
"Add start_time and end_time fields in audits table. User can set the start "
|
||||
"time and/or end time when creating CONTINUOUS audit."
|
||||
@@ -93,6 +186,19 @@ msgstr ""
|
||||
"Add superseded state for an action plan if the cluster data model has "
|
||||
"changed after it has been created."
|
||||
|
||||
msgid ""
|
||||
"Added Placement API helper to Watcher. Now Watcher can get information about "
|
||||
"resource providers, it can be used for the data model and strategies. Config "
|
||||
"group placement_client with options 'api_version', 'interface' and "
|
||||
"'region_name' is also added. The default values for 'api_version' and "
|
||||
"'interface' are 1.29 and 'public', respectively."
|
||||
msgstr ""
|
||||
"Added Placement API helper to Watcher. Now Watcher can get information about "
|
||||
"resource providers, it can be used for the data model and strategies. Config "
|
||||
"group placement_client with options 'api_version', 'interface' and "
|
||||
"'region_name' is also added. The default values for 'api_version' and "
|
||||
"'interface' are 1.29 and 'public', respectively."
|
||||
|
||||
msgid "Added SUSPENDED audit state"
|
||||
msgstr "Added SUSPENDED audit state"
|
||||
|
||||
@@ -107,6 +213,31 @@ msgstr ""
|
||||
"scoring engine by different Strategies, which improve the code and data "
|
||||
"model re-use."
|
||||
|
||||
msgid ""
|
||||
"Added a new config option 'action_execution_rule' which is a dict type. Its "
|
||||
"key field is strategy name and the value is 'ALWAYS' or 'ANY'. 'ALWAYS' "
|
||||
"means the callback function returns True as usual. 'ANY' means the return "
|
||||
"depends on the result of previous action execution. The callback returns "
|
||||
"True if previous action gets failed, and the engine continues to run the "
|
||||
"next action. If previous action executes success, the callback returns False "
|
||||
"then the next action will be ignored. For strategies that aren't in "
|
||||
"'action_execution_rule', the callback always returns True. Please add the "
|
||||
"next section in the watcher.conf file if your strategy needs this feature. "
|
||||
"[watcher_workflow_engines.taskflow] action_execution_rule = {'your strategy "
|
||||
"name': 'ANY'}"
|
||||
msgstr ""
|
||||
"Added a new config option 'action_execution_rule' which is a dict type. Its "
|
||||
"key field is strategy name and the value is 'ALWAYS' or 'ANY'. 'ALWAYS' "
|
||||
"means the callback function returns True as usual. 'ANY' means the return "
|
||||
"depends on the result of previous action execution. The callback returns "
|
||||
"True if previous action gets failed, and the engine continues to run the "
|
||||
"next action. If previous action executes success, the callback returns False "
|
||||
"then the next action will be ignored. For strategies that aren't in "
|
||||
"'action_execution_rule', the callback always returns True. Please add the "
|
||||
"next section in the watcher.conf file if your strategy needs this feature. "
|
||||
"[watcher_workflow_engines.taskflow] action_execution_rule = {'your strategy "
|
||||
"name': 'ANY'}"
|
||||
|
||||
msgid ""
|
||||
"Added a new strategy based on the airflow of servers. This strategy makes "
|
||||
"decisions to migrate VMs to make the airflow uniform."
|
||||
@@ -248,6 +379,15 @@ msgstr ""
|
||||
"The strategy migrates many instances and volumes efficiently with minimum "
|
||||
"downtime automatically."
|
||||
|
||||
msgid ""
|
||||
"Added strategy \"node resource consolidation\". This strategy is used to "
|
||||
"centralize VMs to as few nodes as possible by VM migration. User can set an "
|
||||
"input parameter to decide how to select the destination node."
|
||||
msgstr ""
|
||||
"Added strategy \"node resource consolidation\". This strategy is used to "
|
||||
"centralize VMs to as few nodes as possible by VM migration. User can set an "
|
||||
"input parameter to decide how to select the destination node."
|
||||
|
||||
msgid ""
|
||||
"Added strategy to identify and migrate a Noisy Neighbor - a low priority VM "
|
||||
"that negatively affects peformance of a high priority VM by over utilizing "
|
||||
@@ -284,6 +424,19 @@ msgstr ""
|
||||
msgid "Adds baremetal data model in Watcher"
|
||||
msgstr "Adds baremetal data model in Watcher"
|
||||
|
||||
msgid ""
|
||||
"All datasources can now be configured to retry retrieving a metric upon "
|
||||
"encountering an error. Between each attempt will be a set amount of time "
|
||||
"which can be adjusted from the configuration. These configuration options "
|
||||
"can be found in the `[watcher_datasources]` group and are named "
|
||||
"`query_max_retries` and `query_timeout`."
|
||||
msgstr ""
|
||||
"All datasources can now be configured to retry retrieving a metric upon "
|
||||
"encountering an error. Between each attempt will be a set amount of time "
|
||||
"which can be adjusted from the configuration. These configuration options "
|
||||
"can be found in the `[watcher_datasources]` group and are named "
|
||||
"`query_max_retries` and `query_timeout`."
|
||||
|
||||
msgid ""
|
||||
"Allow decision engine to pass strategy parameters, like optimization "
|
||||
"threshold, to selected strategy, also strategy to provide parameters info to "
|
||||
@@ -293,6 +446,34 @@ msgstr ""
|
||||
"threshold, to selected strategy, also strategy to provide parameters info to "
|
||||
"end user."
|
||||
|
||||
msgid ""
|
||||
"Allow using file to override metric map. Override the metric map of each "
|
||||
"datasource as soon as it is created by the manager. This override comes from "
|
||||
"a file whose path is provided by a setting in config file. The setting is "
|
||||
"`watcher_decision_engine/metric_map_path`. The file contains a map per "
|
||||
"datasource whose keys are the metric names as recognized by watcher and the "
|
||||
"value is the real name of the metric in the datasource. This setting "
|
||||
"defaults to `/etc/watcher/metric_map.yaml`, and presence of this file is "
|
||||
"optional."
|
||||
msgstr ""
|
||||
"Allow using file to override metric map. Override the metric map of each "
|
||||
"datasource as soon as it is created by the manager. This override comes from "
|
||||
"a file whose path is provided by a setting in config file. The setting is "
|
||||
"`watcher_decision_engine/metric_map_path`. The file contains a map per "
|
||||
"datasource whose keys are the metric names as recognized by watcher and the "
|
||||
"value is the real name of the metric in the datasource. This setting "
|
||||
"defaults to `/etc/watcher/metric_map.yaml`, and presence of this file is "
|
||||
"optional."
|
||||
|
||||
msgid ""
|
||||
"An Watcher API WSGI application script ``watcher-api-wsgi`` is now "
|
||||
"available. It is auto-generated by ``pbr`` and allows to run the API service "
|
||||
"using WSGI server (for example Nginx and uWSGI)."
|
||||
msgstr ""
|
||||
"An Watcher API WSGI application script ``watcher-api-wsgi`` is now "
|
||||
"available. It is auto-generated by ``pbr`` and allows to run the API service "
|
||||
"using WSGI server (for example Nginx and uWSGI)."
|
||||
|
||||
msgid ""
|
||||
"Audits have 'name' field now, that is more friendly to end users. Audit's "
|
||||
"name can't exceed 63 characters."
|
||||
@@ -300,9 +481,25 @@ msgstr ""
|
||||
"Audits have 'name' field now, that is more friendly to end users. Audit's "
|
||||
"name can't exceed 63 characters."
|
||||
|
||||
msgid ""
|
||||
"Baremetal Model gets Audit scoper with an ability to exclude Ironic nodes."
|
||||
msgstr ""
|
||||
"Baremetal Model gets Audit scope with an ability to exclude Ironic nodes."
|
||||
|
||||
msgid "Bug Fixes"
|
||||
msgstr "Bug Fixes"
|
||||
|
||||
msgid ""
|
||||
"Ceilometer Datasource has been deprecated since its API has been deprecated "
|
||||
"in Ocata cycle. Watcher has supported Ceilometer for some releases after "
|
||||
"Ocata to let users migrate to Gnocchi/Monasca datasources. Since Train "
|
||||
"release, Ceilometer support will be removed."
|
||||
msgstr ""
|
||||
"Ceilometer Datasource has been deprecated since its API has been deprecated "
|
||||
"in Ocata cycle. Watcher has supported Ceilometer for some releases after "
|
||||
"Ocata to let users migrate to Gnocchi/Monasca datasources. Since Train "
|
||||
"release, Ceilometer support will be removed."
|
||||
|
||||
msgid "Centralize all configuration options for Watcher."
|
||||
msgstr "Centralise all configuration options for Watcher."
|
||||
|
||||
@@ -360,6 +557,52 @@ msgstr ""
|
||||
"Now instances from particular project in OpenStack can be excluded from "
|
||||
"audit defining scope in audit templates."
|
||||
|
||||
msgid ""
|
||||
"For a large cloud infrastructure, retrieving data from Nova may take a long "
|
||||
"time. To avoid getting too much data from Nova, building the compute data "
|
||||
"model according to the scope of audit."
|
||||
msgstr ""
|
||||
"For a large cloud infrastructure, retrieving data from Nova may take a long "
|
||||
"time. To avoid getting too much data from Nova, building the compute data "
|
||||
"model according to the scope of audit."
|
||||
|
||||
msgid ""
|
||||
"Grafana has been added as datasource that can be used for collecting "
|
||||
"metrics. The configuration options allow to specify what metrics and how "
|
||||
"they are stored in grafana so that no matter how Grafana is configured it "
|
||||
"can still be used. The configuration can be done via the typical "
|
||||
"configuration file but it is recommended to configure most options in the "
|
||||
"yaml file for metrics. For a complete walkthrough on configuring Grafana "
|
||||
"see: https://docs.openstack.org/watcher/latest/datasources/grafana.html"
|
||||
msgstr ""
|
||||
"Grafana has been added as datasource that can be used for collecting "
|
||||
"metrics. The configuration options allow to specify what metrics and how "
|
||||
"they are stored in Grafana so that no matter how Grafana is configured it "
|
||||
"can still be used. The configuration can be done via the typical "
|
||||
"configuration file but it is recommended to configure most options in the "
|
||||
"yaml file for metrics. For a complete walkthrough on configuring Grafana "
|
||||
"see: https://docs.openstack.org/watcher/latest/datasources/grafana.html"
|
||||
|
||||
msgid ""
|
||||
"If Gnocchi was configured to have a custom amount of retries and or a custom "
|
||||
"timeout then the configuration needs to moved into the "
|
||||
"`[watcher_datasources]` group instead of the `[gnocchi_client]` group."
|
||||
msgstr ""
|
||||
"If Gnocchi was configured to have a custom amount of retries and or a custom "
|
||||
"timeout then the configuration needs to moved into the "
|
||||
"`[watcher_datasources]` group instead of the `[gnocchi_client]` group."
|
||||
|
||||
msgid ""
|
||||
"Improved interface for datasource baseclass that better defines expected "
|
||||
"values and types for parameters and return types of all abstract methods. "
|
||||
"This allows all strategies to work with every datasource provided the "
|
||||
"metrics are configured for that given datasource."
|
||||
msgstr ""
|
||||
"Improved interface for datasource baseclass that better defines expected "
|
||||
"values and types for parameters and return types of all abstract methods. "
|
||||
"This allows all strategies to work with every datasource provided the "
|
||||
"metrics are configured for that given datasource."
|
||||
|
||||
msgid ""
|
||||
"Instance cold migration logic is now replaced with using Nova migrate "
|
||||
"Server(migrate Action) API which has host option since v2.56."
|
||||
@@ -367,6 +610,17 @@ msgstr ""
|
||||
"Instance cold migration logic is now replaced with using Nova migrate "
|
||||
"Server(migrate Action) API which has host option since v2.56."
|
||||
|
||||
msgid ""
|
||||
"Many operations in the decision engine will block on I/O. Such I/O "
|
||||
"operations can stall the execution of a sequential application "
|
||||
"significantly. To reduce the potential bottleneck of many operations the "
|
||||
"general purpose decision engine threadpool is introduced."
|
||||
msgstr ""
|
||||
"Many operations in the decision engine will block on I/O. Such I/O "
|
||||
"operations can stall the execution of a sequential application "
|
||||
"significantly. To reduce the potential bottleneck of many operations the "
|
||||
"general purpose decision engine threadpool is introduced."
|
||||
|
||||
msgid "New Features"
|
||||
msgstr "New Features"
|
||||
|
||||
@@ -389,6 +643,13 @@ msgstr ""
|
||||
"Nova API version is now set to 2.56 by default. This needs the migrate "
|
||||
"action of migration type cold with destination_node parameter to work."
|
||||
|
||||
msgid ""
|
||||
"Now Watcher strategy can select specific planner beyond default. Strategy "
|
||||
"can set planner property to specify its own planner."
|
||||
msgstr ""
|
||||
"Now Watcher strategy can select specific planner beyond default. Strategy "
|
||||
"can set planner property to specify its own planner."
|
||||
|
||||
msgid "Ocata Series Release Notes"
|
||||
msgstr "Ocata Series Release Notes"
|
||||
|
||||
@@ -429,12 +690,60 @@ msgstr ""
|
||||
"resources will be called \"Audit scope\" and will be defined in each audit "
|
||||
"template (which contains the audit settings)."
|
||||
|
||||
msgid ""
|
||||
"Python 2.7 support has been dropped. Last release of Watcher to support "
|
||||
"py2.7 is OpenStack Train. The minimum version of Python now supported by "
|
||||
"Watcher is Python 3.6."
|
||||
msgstr ""
|
||||
"Python 2.7 support has been dropped. Last release of Watcher to support "
|
||||
"py2.7 is OpenStack Train. The minimum version of Python now supported by "
|
||||
"Watcher is Python 3.6."
|
||||
|
||||
msgid "Queens Series Release Notes"
|
||||
msgstr "Queens Series Release Notes"
|
||||
|
||||
msgid "Rocky Series Release Notes"
|
||||
msgstr "Rocky Series Release Notes"
|
||||
|
||||
msgid ""
|
||||
"Several strategies have changed the `node` parameter to `compute_node` to be "
|
||||
"better aligned with terminology. These strategies include "
|
||||
"`basic_consolidation` and `workload_stabilzation`. The `node` parameter will "
|
||||
"remain supported during Train release and will be removed in the subsequent "
|
||||
"release."
|
||||
msgstr ""
|
||||
"Several strategies have changed the `node` parameter to `compute_node` to be "
|
||||
"better aligned with terminology. These strategies include "
|
||||
"`basic_consolidation` and `workload_stabilzation`. The `node` parameter will "
|
||||
"remain supported during Train release and will be removed in the subsequent "
|
||||
"release."
|
||||
|
||||
msgid ""
|
||||
"Specific strategies can override this order and use datasources which are "
|
||||
"not listed in the global preference."
|
||||
msgstr ""
|
||||
"Specific strategies can override this order and use datasources which are "
|
||||
"not listed in the global preference."
|
||||
|
||||
msgid "Stein Series Release Notes"
|
||||
msgstr "Stein Series Release Notes"
|
||||
|
||||
msgid ""
|
||||
"The building of the compute (Nova) data model will be done using the "
|
||||
"decision engine threadpool, thereby, significantly reducing the total time "
|
||||
"required to build it."
|
||||
msgstr ""
|
||||
"The building of the compute (Nova) data model will be done using the "
|
||||
"decision engine threadpool, thereby, significantly reducing the total time "
|
||||
"required to build it."
|
||||
|
||||
msgid ""
|
||||
"The configuration options for query retries in `[gnocchi_client]` are "
|
||||
"deprecated and the option in `[watcher_datasources]` should now be used."
|
||||
msgstr ""
|
||||
"The configuration options for query retries in `[gnocchi_client]` are "
|
||||
"deprecated and the option in `[watcher_datasources]` should now be used."
|
||||
|
||||
msgid ""
|
||||
"The graph model describes how VMs are associated to compute hosts. This "
|
||||
"allows for seeing relationships upfront between the entities and hence can "
|
||||
@@ -455,6 +764,22 @@ msgstr ""
|
||||
"was fixed. Before fixing, it booted an instance in the service project as a "
|
||||
"migrated instance."
|
||||
|
||||
msgid ""
|
||||
"The minimum required version of the ``[nova_client]/api_version`` value is "
|
||||
"now enforced to be ``2.56`` which is available since the Queens version of "
|
||||
"the nova compute service."
|
||||
msgstr ""
|
||||
"The minimum required version of the ``[nova_client]/api_version`` value is "
|
||||
"now enforced to be ``2.56`` which is available since the Queens version of "
|
||||
"the Nova compute service."
|
||||
|
||||
msgid ""
|
||||
"The new strategy baseclass has significant changes in method parameters and "
|
||||
"any out-of-tree strategies will have to be adopted."
|
||||
msgstr ""
|
||||
"The new strategy baseclass has significant changes in method parameters and "
|
||||
"any out-of-tree strategies will have to be adopted."
|
||||
|
||||
msgid ""
|
||||
"There is new ability to create Watcher continuous audits with cron interval. "
|
||||
"It means you may use, for example, optional argument '--interval \"\\*/5 \\* "
|
||||
@@ -468,9 +793,27 @@ msgstr ""
|
||||
"best effort basis and therefore, we recommend you to use a minimal cron "
|
||||
"interval of at least one minute."
|
||||
|
||||
msgid "Train Series Release Notes"
|
||||
msgstr "Train Series Release Notes"
|
||||
|
||||
msgid "Upgrade Notes"
|
||||
msgstr "Upgrade Notes"
|
||||
|
||||
msgid ""
|
||||
"Using ``watcher/api/app.wsgi`` script is deprecated and it will be removed "
|
||||
"in U release. Please switch to automatically generated ``watcher-api-wsgi`` "
|
||||
"script instead."
|
||||
msgstr ""
|
||||
"Using ``watcher/api/app.wsgi`` script is deprecated and it will be removed "
|
||||
"in U release. Please switch to automatically generated ``watcher-api-wsgi`` "
|
||||
"script instead."
|
||||
|
||||
msgid "Ussuri Series Release Notes"
|
||||
msgstr "Ussuri Series Release Notes"
|
||||
|
||||
msgid "Victoria Series Release Notes"
|
||||
msgstr "Victoria Series Release Notes"
|
||||
|
||||
msgid ""
|
||||
"Watcher can continuously optimize the OpenStack cloud for a specific "
|
||||
"strategy or goal by triggering an audit periodically which generates an "
|
||||
@@ -480,6 +823,15 @@ msgstr ""
|
||||
"strategy or goal by triggering an audit periodically which generates an "
|
||||
"action plan and run it automatically."
|
||||
|
||||
msgid ""
|
||||
"Watcher can get resource information such as total, allocation ratio and "
|
||||
"reserved information from Placement API. Now we add some new fields to the "
|
||||
"Watcher Data Model:"
|
||||
msgstr ""
|
||||
"Watcher can get resource information such as total, allocation ratio and "
|
||||
"reserved information from Placement API. Now we add some new fields to the "
|
||||
"Watcher Data Model:"
|
||||
|
||||
msgid ""
|
||||
"Watcher can now run specific actions in parallel improving the performances "
|
||||
"dramatically when executing an action plan."
|
||||
@@ -517,6 +869,15 @@ msgstr ""
|
||||
"includes all instances. It filters excluded instances when migration during "
|
||||
"the audit."
|
||||
|
||||
msgid ""
|
||||
"Watcher now supports configuring which datasource to use and in which order. "
|
||||
"This configuration is done by specifying datasources in the "
|
||||
"watcher_datasources section:"
|
||||
msgstr ""
|
||||
"Watcher now supports configuring which datasource to use and in which order. "
|
||||
"This configuration is done by specifying datasources in the "
|
||||
"watcher_datasources section:"
|
||||
|
||||
msgid ""
|
||||
"Watcher removes the support to Nova legacy notifications because of Nova "
|
||||
"will deprecate them."
|
||||
@@ -557,9 +918,15 @@ msgstr ""
|
||||
"Watcher supports multiple metrics backend and relies on Ceilometer and "
|
||||
"Monasca."
|
||||
|
||||
msgid "We also add some new propeties:"
|
||||
msgstr "We also add some new properties:"
|
||||
|
||||
msgid "Welcome to watcher's Release Notes documentation!"
|
||||
msgstr "Welcome to watcher's Release Notes documentation!"
|
||||
|
||||
msgid "``[watcher_datasources] datasources = gnocchi,monasca,ceilometer``"
|
||||
msgstr "``[watcher_datasources] datasources = gnocchi,monasca,ceilometer``"
|
||||
|
||||
msgid ""
|
||||
"all Watcher objects have been refactored to support OVO (oslo."
|
||||
"versionedobjects) which was a prerequisite step in order to implement "
|
||||
@@ -569,6 +936,21 @@ msgstr ""
|
||||
"versionedobjects) which was a prerequisite step in order to implement "
|
||||
"versioned notifications."
|
||||
|
||||
msgid ""
|
||||
"disk_gb_capacity: The amount of disk, take allocation ratio into account, "
|
||||
"but do not include reserved."
|
||||
msgstr ""
|
||||
"disk_gb_capacity: The amount of disk, take allocation ratio into account, "
|
||||
"but do not include reserved."
|
||||
|
||||
msgid ""
|
||||
"disk_gb_reserved: The amount of disk a node has reserved for its own use."
|
||||
msgstr ""
|
||||
"disk_gb_reserved: The amount of disk a node has reserved for its own use."
|
||||
|
||||
msgid "disk_ratio: Disk allocation ratio."
|
||||
msgstr "disk_ratio: Disk allocation ratio."
|
||||
|
||||
msgid "instance.create.end"
|
||||
msgstr "instance.create.end"
|
||||
|
||||
@@ -635,6 +1017,21 @@ msgstr "instance.unshelve.end"
|
||||
msgid "instance.update"
|
||||
msgstr "instance.update"
|
||||
|
||||
msgid ""
|
||||
"memory_mb_capacity: The amount of memory, take allocation ratio into "
|
||||
"account, but do not include reserved."
|
||||
msgstr ""
|
||||
"memory_mb_capacity: The amount of memory, take allocation ratio into "
|
||||
"account, but do not include reserved."
|
||||
|
||||
msgid ""
|
||||
"memory_mb_reserved: The amount of memory a node has reserved for its own use."
|
||||
msgstr ""
|
||||
"memory_mb_reserved: The amount of memory a node has reserved for its own use."
|
||||
|
||||
msgid "memory_ratio: Memory allocation ratio."
|
||||
msgstr "memory_ratio: Memory allocation ratio."
|
||||
|
||||
msgid "new:"
|
||||
msgstr "new:"
|
||||
|
||||
@@ -649,3 +1046,16 @@ msgstr "service.delete"
|
||||
|
||||
msgid "service.update"
|
||||
msgstr "service.update"
|
||||
|
||||
msgid ""
|
||||
"vcpu_capacity: The amount of vcpu, take allocation ratio into account, but "
|
||||
"do not include reserved."
|
||||
msgstr ""
|
||||
"vcpu_capacity: The amount of vcpu, take allocation ratio into account, but "
|
||||
"do not include reserved."
|
||||
|
||||
msgid "vcpu_ratio: CPU allocation ratio."
|
||||
msgstr "vcpu_ratio: CPU allocation ratio."
|
||||
|
||||
msgid "vcpu_reserved: The amount of cpu a node has reserved for its own use."
|
||||
msgstr "vcpu_reserved: The amount of CPU a node has reserved for its own use."
|
||||
|
||||
6
releasenotes/source/ussuri.rst
Normal file
6
releasenotes/source/ussuri.rst
Normal file
@@ -0,0 +1,6 @@
|
||||
===========================
|
||||
Ussuri Series Release Notes
|
||||
===========================
|
||||
|
||||
.. release-notes::
|
||||
:branch: stable/ussuri
|
||||
6
releasenotes/source/victoria.rst
Normal file
6
releasenotes/source/victoria.rst
Normal file
@@ -0,0 +1,6 @@
|
||||
=============================
|
||||
Victoria Series Release Notes
|
||||
=============================
|
||||
|
||||
.. release-notes::
|
||||
:branch: stable/victoria
|
||||
@@ -1,34 +1,34 @@
|
||||
# The order of packages is significant, because pip processes them in the order
|
||||
# The order of packages is significant, because pip processes them in the order
|
||||
# of appearance. Changing the order has an impact on the overall integration
|
||||
# process, which may cause wedges in the gate later.
|
||||
|
||||
apscheduler>=3.5.1 # MIT License
|
||||
jsonpatch>=1.21 # BSD
|
||||
keystoneauth1>=3.4.0 # Apache-2.0
|
||||
jsonschema>=2.6.0 # MIT
|
||||
jsonschema>=3.2.0 # MIT
|
||||
keystonemiddleware>=4.21.0 # Apache-2.0
|
||||
lxml>=4.1.1 # BSD
|
||||
lxml>=4.5.1 # BSD
|
||||
croniter>=0.3.20 # MIT License
|
||||
os-resource-classes>=0.4.0
|
||||
oslo.concurrency>=3.26.0 # Apache-2.0
|
||||
oslo.cache>=1.29.0 # Apache-2.0
|
||||
oslo.config>=5.2.0 # Apache-2.0
|
||||
oslo.config>=6.8.0 # Apache-2.0
|
||||
oslo.context>=2.21.0 # Apache-2.0
|
||||
oslo.db>=4.35.0 # Apache-2.0
|
||||
oslo.db>=4.44.0 # Apache-2.0
|
||||
oslo.i18n>=3.20.0 # Apache-2.0
|
||||
oslo.log>=3.37.0 # Apache-2.0
|
||||
oslo.messaging>=8.1.2 # Apache-2.0
|
||||
oslo.policy>=1.34.0 # Apache-2.0
|
||||
oslo.policy>=3.6.0 # Apache-2.0
|
||||
oslo.reports>=1.27.0 # Apache-2.0
|
||||
oslo.serialization>=2.25.0 # Apache-2.0
|
||||
oslo.service>=1.30.0 # Apache-2.0
|
||||
oslo.upgradecheck>=0.1.0 # Apache-2.0
|
||||
oslo.upgradecheck>=1.3.0 # Apache-2.0
|
||||
oslo.utils>=3.36.0 # Apache-2.0
|
||||
oslo.versionedobjects>=1.32.0 # Apache-2.0
|
||||
PasteDeploy>=1.5.2 # MIT
|
||||
pbr>=3.1.1 # Apache-2.0
|
||||
pecan>=1.3.2 # BSD
|
||||
PrettyTable<0.8,>=0.7.2 # BSD
|
||||
PrettyTable>=0.7.2 # BSD
|
||||
gnocchiclient>=7.0.1 # Apache-2.0
|
||||
python-ceilometerclient>=2.9.0 # Apache-2.0
|
||||
python-cinderclient>=3.5.0 # Apache-2.0
|
||||
@@ -39,12 +39,11 @@ python-neutronclient>=6.7.0 # Apache-2.0
|
||||
python-novaclient>=14.1.0 # Apache-2.0
|
||||
python-openstackclient>=3.14.0 # Apache-2.0
|
||||
python-ironicclient>=2.5.0 # Apache-2.0
|
||||
six>=1.11.0 # MIT
|
||||
SQLAlchemy>=1.2.5 # MIT
|
||||
stevedore>=1.28.0 # Apache-2.0
|
||||
taskflow>=3.7.1 # Apache-2.0
|
||||
taskflow>=3.8.0 # Apache-2.0
|
||||
WebOb>=1.8.5 # MIT
|
||||
WSME>=0.9.2 # MIT
|
||||
networkx>=2.2;python_version>='3.4' # BSD
|
||||
networkx>=2.4 # BSD
|
||||
microversion_parse>=0.2.1 # Apache-2.0
|
||||
futurist>=1.8.0 # Apache-2.0
|
||||
|
||||
16
setup.cfg
16
setup.cfg
@@ -19,6 +19,7 @@ classifier =
|
||||
Programming Language :: Python :: 3
|
||||
Programming Language :: Python :: 3.6
|
||||
Programming Language :: Python :: 3.7
|
||||
Programming Language :: Python :: 3.8
|
||||
|
||||
[files]
|
||||
packages =
|
||||
@@ -107,18 +108,3 @@ watcher_cluster_data_model_collectors =
|
||||
compute = watcher.decision_engine.model.collector.nova:NovaClusterDataModelCollector
|
||||
storage = watcher.decision_engine.model.collector.cinder:CinderClusterDataModelCollector
|
||||
baremetal = watcher.decision_engine.model.collector.ironic:BaremetalClusterDataModelCollector
|
||||
|
||||
|
||||
[compile_catalog]
|
||||
directory = watcher/locale
|
||||
domain = watcher
|
||||
|
||||
[update_catalog]
|
||||
domain = watcher
|
||||
output_dir = watcher/locale
|
||||
input_file = watcher/locale/watcher.pot
|
||||
|
||||
[extract_messages]
|
||||
keywords = _ gettext ngettext l_ lazy_gettext _LI _LW _LE _LC
|
||||
mapping_file = babel.cfg
|
||||
output_file = watcher/locale/watcher.pot
|
||||
|
||||
@@ -5,8 +5,7 @@
|
||||
coverage>=4.5.1 # Apache-2.0
|
||||
doc8>=0.8.0 # Apache-2.0
|
||||
freezegun>=0.3.10 # Apache-2.0
|
||||
hacking>=3.0,<3.1.0 # Apache-2.0
|
||||
mock>=2.0.0 # BSD
|
||||
hacking>=3.0.1,<3.1.0 # Apache-2.0
|
||||
oslotest>=3.3.0 # Apache-2.0
|
||||
os-testr>=1.0.0 # Apache-2.0
|
||||
testscenarios>=0.5.0 # Apache-2.0/BSD
|
||||
|
||||
17
tox.ini
17
tox.ini
@@ -1,6 +1,6 @@
|
||||
[tox]
|
||||
minversion = 2.0
|
||||
envlist = py36,py37,pep8
|
||||
envlist = py38,pep8
|
||||
skipsdist = True
|
||||
ignore_basepython_conflict = True
|
||||
|
||||
@@ -13,7 +13,7 @@ install_command = pip install {opts} {packages}
|
||||
setenv =
|
||||
VIRTUAL_ENV={envdir}
|
||||
deps =
|
||||
-c{env:UPPER_CONSTRAINTS_FILE:https://releases.openstack.org/constraints/upper/master}
|
||||
-c{env:TOX_CONSTRAINTS_FILE:https://releases.openstack.org/constraints/upper/wallaby}
|
||||
-r{toxinidir}/test-requirements.txt
|
||||
-r{toxinidir}/requirements.txt
|
||||
commands =
|
||||
@@ -31,7 +31,7 @@ commands =
|
||||
[testenv:venv]
|
||||
setenv = PYTHONHASHSEED=0
|
||||
deps =
|
||||
-c{env:UPPER_CONSTRAINTS_FILE:https://releases.openstack.org/constraints/upper/master}
|
||||
-c{env:TOX_CONSTRAINTS_FILE:https://releases.openstack.org/constraints/upper/wallaby}
|
||||
-r{toxinidir}/doc/requirements.txt
|
||||
-r{toxinidir}/test-requirements.txt
|
||||
-r{toxinidir}/requirements.txt
|
||||
@@ -49,7 +49,9 @@ commands =
|
||||
|
||||
[testenv:docs]
|
||||
setenv = PYTHONHASHSEED=0
|
||||
deps = -r{toxinidir}/doc/requirements.txt
|
||||
deps =
|
||||
-c{env:UPPER_CONSTRAINTS_FILE:https://releases.openstack.org/constraints/upper/master}
|
||||
-r{toxinidir}/doc/requirements.txt
|
||||
commands =
|
||||
rm -fr doc/build doc/source/api/ .autogenerated
|
||||
sphinx-build -W --keep-going -b html doc/source doc/build/html
|
||||
@@ -105,6 +107,7 @@ extension =
|
||||
N340 = checks:check_oslo_i18n_wrapper
|
||||
N341 = checks:check_builtins_gettext
|
||||
N342 = checks:no_redundant_import_alias
|
||||
N366 = checks:import_stock_mock
|
||||
paths = ./watcher/hacking
|
||||
|
||||
|
||||
@@ -131,9 +134,3 @@ commands = sphinx-build -a -W -E -d releasenotes/build/doctrees --keep-going -b
|
||||
[testenv:bandit]
|
||||
deps = -r{toxinidir}/test-requirements.txt
|
||||
commands = bandit -r watcher -x watcher/tests/* -n5 -ll -s B320
|
||||
|
||||
[testenv:lower-constraints]
|
||||
deps =
|
||||
-c{toxinidir}/lower-constraints.txt
|
||||
-r{toxinidir}/test-requirements.txt
|
||||
-r{toxinidir}/requirements.txt
|
||||
|
||||
@@ -13,8 +13,6 @@
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
from __future__ import unicode_literals
|
||||
|
||||
from oslo_config import cfg
|
||||
from watcher.api import hooks
|
||||
|
||||
|
||||
@@ -138,6 +138,9 @@ class AuditTemplatePostType(wtypes.Base):
|
||||
raise exception.InvalidGoal(goal=audit_template.goal)
|
||||
|
||||
if audit_template.scope:
|
||||
keys = [list(s)[0] for s in audit_template.scope]
|
||||
if keys[0] not in ('compute', 'storage'):
|
||||
audit_template.scope = [dict(compute=audit_template.scope)]
|
||||
common_utils.Draft4Validator(
|
||||
AuditTemplatePostType._build_schema()
|
||||
).validate(audit_template.scope)
|
||||
@@ -158,18 +161,23 @@ class AuditTemplatePostType(wtypes.Base):
|
||||
"included and excluded together"))
|
||||
|
||||
if audit_template.strategy:
|
||||
available_strategies = objects.Strategy.list(
|
||||
AuditTemplatePostType._ctx)
|
||||
available_strategies_map = {
|
||||
s.uuid: s for s in available_strategies}
|
||||
if audit_template.strategy not in available_strategies_map:
|
||||
try:
|
||||
if (common_utils.is_uuid_like(audit_template.strategy) or
|
||||
common_utils.is_int_like(audit_template.strategy)):
|
||||
strategy = objects.Strategy.get(
|
||||
AuditTemplatePostType._ctx, audit_template.strategy)
|
||||
else:
|
||||
strategy = objects.Strategy.get_by_name(
|
||||
AuditTemplatePostType._ctx, audit_template.strategy)
|
||||
except Exception:
|
||||
raise exception.InvalidStrategy(
|
||||
strategy=audit_template.strategy)
|
||||
|
||||
strategy = available_strategies_map[audit_template.strategy]
|
||||
# Check that the strategy we indicate is actually related to the
|
||||
# specified goal
|
||||
if strategy.goal_id != goal.id:
|
||||
available_strategies = objects.Strategy.list(
|
||||
AuditTemplatePostType._ctx)
|
||||
choices = ["'%s' (%s)" % (s.uuid, s.name)
|
||||
for s in available_strategies]
|
||||
raise exception.InvalidStrategy(
|
||||
|
||||
@@ -19,8 +19,6 @@ Service mechanism provides ability to monitor Watcher services state.
|
||||
"""
|
||||
|
||||
import datetime
|
||||
import six
|
||||
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
from oslo_utils import timeutils
|
||||
@@ -70,7 +68,7 @@ class Service(base.APIBase):
|
||||
service = objects.Service.get(pecan.request.context, id)
|
||||
last_heartbeat = (service.last_seen_up or service.updated_at or
|
||||
service.created_at)
|
||||
if isinstance(last_heartbeat, six.string_types):
|
||||
if isinstance(last_heartbeat, str):
|
||||
# NOTE(russellb) If this service came in over rpc via
|
||||
# conductor, then the timestamp will be a string and needs to be
|
||||
# converted back to a datetime.
|
||||
|
||||
@@ -15,7 +15,6 @@
|
||||
|
||||
from oslo_serialization import jsonutils
|
||||
from oslo_utils import strutils
|
||||
import six
|
||||
import wsme
|
||||
from wsme import types as wtypes
|
||||
|
||||
@@ -132,7 +131,7 @@ class JsonType(wtypes.UserType):
|
||||
|
||||
def __str__(self):
|
||||
# These are the json serializable native types
|
||||
return ' | '.join(map(str, (wtypes.text, six.integer_types, float,
|
||||
return ' | '.join(map(str, (wtypes.text, int, float,
|
||||
BooleanType, list, dict, None)))
|
||||
|
||||
@staticmethod
|
||||
|
||||
@@ -15,9 +15,9 @@
|
||||
# under the License.
|
||||
|
||||
|
||||
from http import client as http_client
|
||||
from oslo_config import cfg
|
||||
from pecan import hooks
|
||||
from six.moves import http_client
|
||||
|
||||
from watcher.common import context
|
||||
|
||||
|
||||
@@ -24,7 +24,6 @@ from xml import etree as et
|
||||
|
||||
from oslo_log import log
|
||||
from oslo_serialization import jsonutils
|
||||
import six
|
||||
import webob
|
||||
|
||||
from watcher._i18n import _
|
||||
@@ -84,12 +83,10 @@ class ParsableErrorMiddleware(object):
|
||||
'</error_message>' % state['status_code']]
|
||||
state['headers'].append(('Content-Type', 'application/xml'))
|
||||
else:
|
||||
if six.PY3:
|
||||
app_iter = [i.decode('utf-8') for i in app_iter]
|
||||
app_iter = [i.decode('utf-8') for i in app_iter]
|
||||
body = [jsonutils.dumps(
|
||||
{'error_message': '\n'.join(app_iter)})]
|
||||
if six.PY3:
|
||||
body = [item.encode('utf-8') for item in body]
|
||||
body = [item.encode('utf-8') for item in body]
|
||||
state['headers'].append(('Content-Type', 'application/json'))
|
||||
state['headers'].append(('Content-Length', str(len(body[0]))))
|
||||
else:
|
||||
|
||||
@@ -20,7 +20,6 @@ import itertools
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
from oslo_utils import timeutils
|
||||
import six
|
||||
|
||||
from watcher.common import context as watcher_context
|
||||
from watcher.common import scheduling
|
||||
@@ -83,7 +82,7 @@ class APISchedulingService(scheduling.BackgroundSchedulerService):
|
||||
service = objects.Service.get(context, service_id)
|
||||
last_heartbeat = (service.last_seen_up or service.updated_at or
|
||||
service.created_at)
|
||||
if isinstance(last_heartbeat, six.string_types):
|
||||
if isinstance(last_heartbeat, str):
|
||||
# NOTE(russellb) If this service came in over rpc via
|
||||
# conductor, then the timestamp will be a string and needs to be
|
||||
# converted back to a datetime.
|
||||
|
||||
@@ -18,11 +18,9 @@
|
||||
#
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseActionPlanHandler(object):
|
||||
class BaseActionPlanHandler(object, metaclass=abc.ABCMeta):
|
||||
@abc.abstractmethod
|
||||
def execute(self):
|
||||
raise NotImplementedError()
|
||||
|
||||
@@ -19,14 +19,12 @@
|
||||
import abc
|
||||
|
||||
import jsonschema
|
||||
import six
|
||||
|
||||
from watcher.common import clients
|
||||
from watcher.common.loader import loadable
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseAction(loadable.Loadable):
|
||||
class BaseAction(loadable.Loadable, metaclass=abc.ABCMeta):
|
||||
# NOTE(jed): by convention we decided
|
||||
# that the attribute "resource_id" is the unique id of
|
||||
# the resource to which the Action applies to allow us to use it in the
|
||||
|
||||
@@ -15,8 +15,6 @@
|
||||
# limitations under the License.
|
||||
#
|
||||
|
||||
from __future__ import unicode_literals
|
||||
|
||||
from oslo_log import log
|
||||
|
||||
from watcher.applier.loading import default
|
||||
|
||||
@@ -186,7 +186,7 @@ class Migrate(base.BaseAction):
|
||||
return self.migrate(destination=self.destination_node)
|
||||
|
||||
def revert(self):
|
||||
LOG.info('Migrate action do not revert!')
|
||||
return self.migrate(destination=self.source_node)
|
||||
|
||||
def abort(self):
|
||||
nova = nova_helper.NovaHelper(osc=self.osc)
|
||||
|
||||
@@ -95,7 +95,7 @@ class Resize(base.BaseAction):
|
||||
return self.resize()
|
||||
|
||||
def revert(self):
|
||||
return self.migrate(destination=self.source_node)
|
||||
LOG.warning("revert not supported")
|
||||
|
||||
def pre_condition(self):
|
||||
# TODO(jed): check if the instance exists / check if the instance is on
|
||||
|
||||
@@ -26,11 +26,9 @@ See: :doc:`../architecture` for more details on this component.
|
||||
"""
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseApplier(object):
|
||||
class BaseApplier(object, metaclass=abc.ABCMeta):
|
||||
@abc.abstractmethod
|
||||
def execute(self, action_plan_uuid):
|
||||
raise NotImplementedError()
|
||||
|
||||
@@ -11,9 +11,6 @@
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
from __future__ import unicode_literals
|
||||
|
||||
|
||||
from watcher.common.loader import default
|
||||
|
||||
|
||||
|
||||
@@ -17,7 +17,6 @@
|
||||
#
|
||||
|
||||
import abc
|
||||
import six
|
||||
import time
|
||||
|
||||
import eventlet
|
||||
@@ -40,8 +39,7 @@ CANCEL_STATE = [objects.action_plan.State.CANCELLING,
|
||||
objects.action_plan.State.CANCELLED]
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseWorkFlowEngine(loadable.Loadable):
|
||||
class BaseWorkFlowEngine(loadable.Loadable, metaclass=abc.ABCMeta):
|
||||
|
||||
def __init__(self, config, context=None, applier_manager=None):
|
||||
"""Constructor
|
||||
|
||||
@@ -25,8 +25,11 @@ from taskflow import task as flow_task
|
||||
|
||||
from watcher.applier.workflow_engine import base
|
||||
from watcher.common import exception
|
||||
from watcher import conf
|
||||
from watcher import objects
|
||||
|
||||
CONF = conf.CONF
|
||||
|
||||
LOG = log.getLogger(__name__)
|
||||
|
||||
|
||||
@@ -127,9 +130,11 @@ class DefaultWorkFlowEngine(base.BaseWorkFlowEngine):
|
||||
|
||||
class TaskFlowActionContainer(base.BaseTaskFlowActionContainer):
|
||||
def __init__(self, db_action, engine):
|
||||
name = "action_type:{0} uuid:{1}".format(db_action.action_type,
|
||||
db_action.uuid)
|
||||
super(TaskFlowActionContainer, self).__init__(name, db_action, engine)
|
||||
self.name = "action_type:{0} uuid:{1}".format(db_action.action_type,
|
||||
db_action.uuid)
|
||||
super(TaskFlowActionContainer, self).__init__(self.name,
|
||||
db_action,
|
||||
engine)
|
||||
|
||||
def do_pre_execute(self):
|
||||
db_action = self.engine.notify(self._db_action,
|
||||
@@ -158,6 +163,12 @@ class TaskFlowActionContainer(base.BaseTaskFlowActionContainer):
|
||||
self.action.post_condition()
|
||||
|
||||
def do_revert(self, *args, **kwargs):
|
||||
# NOTE: Not rollback action plan
|
||||
if not CONF.watcher_applier.rollback_when_actionplan_failed:
|
||||
LOG.info("Failed actionplan rollback option is turned off, and "
|
||||
"the following action will be skipped: %s", self.name)
|
||||
return
|
||||
|
||||
LOG.warning("Revert action: %s", self.name)
|
||||
try:
|
||||
# TODO(jed): do we need to update the states in case of failure?
|
||||
|
||||
@@ -18,3 +18,10 @@
|
||||
import eventlet
|
||||
|
||||
eventlet.monkey_patch()
|
||||
|
||||
# Monkey patch the original current_thread to use the up-to-date _active
|
||||
# global variable. See https://bugs.launchpad.net/bugs/1863021 and
|
||||
# https://github.com/eventlet/eventlet/issues/592
|
||||
import __original_module_threading as orig_threading # noqa
|
||||
import threading # noqa
|
||||
orig_threading.current_thread.__globals__['_active'] = threading._active
|
||||
|
||||
@@ -14,6 +14,7 @@
|
||||
|
||||
import sys
|
||||
|
||||
from oslo_upgradecheck import common_checks
|
||||
from oslo_upgradecheck import upgradecheck
|
||||
|
||||
from watcher._i18n import _
|
||||
@@ -43,6 +44,10 @@ class Checks(upgradecheck.UpgradeCommands):
|
||||
_upgrade_checks = (
|
||||
# Added in Train.
|
||||
(_('Minimum Nova API Version'), _minimum_nova_api_version),
|
||||
# Added in Wallaby.
|
||||
(_("Policy File JSON to YAML Migration"),
|
||||
(common_checks.check_policy_json, {'conf': CONF})),
|
||||
|
||||
)
|
||||
|
||||
|
||||
|
||||
@@ -13,7 +13,6 @@
|
||||
from oslo_context import context
|
||||
from oslo_log import log
|
||||
from oslo_utils import timeutils
|
||||
import six
|
||||
|
||||
LOG = log.getLogger(__name__)
|
||||
|
||||
@@ -69,7 +68,7 @@ class RequestContext(context.RequestContext):
|
||||
self.project_id = project_id
|
||||
if not timestamp:
|
||||
timestamp = timeutils.utcnow()
|
||||
if isinstance(timestamp, six.string_types):
|
||||
if isinstance(timestamp, str):
|
||||
timestamp = timeutils.parse_isotime(timestamp)
|
||||
self.timestamp = timestamp
|
||||
self.user_name = user_name
|
||||
|
||||
@@ -14,14 +14,10 @@
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
from __future__ import unicode_literals
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseLoader(object):
|
||||
class BaseLoader(object, metaclass=abc.ABCMeta):
|
||||
|
||||
@abc.abstractmethod
|
||||
def list_available(self):
|
||||
|
||||
@@ -14,8 +14,6 @@
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
from __future__ import unicode_literals
|
||||
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
from stevedore import driver as drivermanager
|
||||
|
||||
@@ -16,13 +16,10 @@
|
||||
|
||||
import abc
|
||||
|
||||
import six
|
||||
|
||||
from watcher.common import service
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class Loadable(object):
|
||||
class Loadable(object, metaclass=abc.ABCMeta):
|
||||
"""Generic interface for dynamically loading a driver/entry point.
|
||||
|
||||
This defines the contract in order to let the loader manager inject
|
||||
@@ -48,8 +45,7 @@ LoadableSingletonMeta = type(
|
||||
"LoadableSingletonMeta", (abc.ABCMeta, service.Singleton), {})
|
||||
|
||||
|
||||
@six.add_metaclass(LoadableSingletonMeta)
|
||||
class LoadableSingleton(object):
|
||||
class LoadableSingleton(object, metaclass=LoadableSingletonMeta):
|
||||
"""Generic interface for dynamically loading a driver as a singleton.
|
||||
|
||||
This defines the contract in order to let the loader manager inject
|
||||
|
||||
@@ -18,6 +18,7 @@
|
||||
import sys
|
||||
|
||||
from oslo_config import cfg
|
||||
from oslo_policy import opts
|
||||
from oslo_policy import policy
|
||||
|
||||
from watcher.common import exception
|
||||
@@ -26,6 +27,12 @@ from watcher.common import policies
|
||||
_ENFORCER = None
|
||||
CONF = cfg.CONF
|
||||
|
||||
# TODO(gmann): Remove setting the default value of config policy_file
|
||||
# once oslo_policy change the default value to 'policy.yaml'.
|
||||
# https://github.com/openstack/oslo.policy/blob/a626ad12fe5a3abd49d70e3e5b95589d279ab578/oslo_policy/opts.py#L49
|
||||
DEFAULT_POLICY_FILE = 'policy.yaml'
|
||||
opts.set_defaults(CONF, DEFAULT_POLICY_FILE)
|
||||
|
||||
|
||||
# we can get a policy enforcer by this init.
|
||||
# oslo policy support change policy rule dynamically.
|
||||
|
||||
@@ -121,22 +121,40 @@ class RequestContextSerializer(messaging.Serializer):
|
||||
def get_client(target, version_cap=None, serializer=None):
|
||||
assert TRANSPORT is not None
|
||||
serializer = RequestContextSerializer(serializer)
|
||||
return messaging.RPCClient(TRANSPORT,
|
||||
target,
|
||||
version_cap=version_cap,
|
||||
serializer=serializer)
|
||||
return messaging.RPCClient(
|
||||
TRANSPORT,
|
||||
target,
|
||||
version_cap=version_cap,
|
||||
serializer=serializer
|
||||
)
|
||||
|
||||
|
||||
def get_server(target, endpoints, serializer=None):
|
||||
assert TRANSPORT is not None
|
||||
access_policy = dispatcher.DefaultRPCAccessPolicy
|
||||
serializer = RequestContextSerializer(serializer)
|
||||
return messaging.get_rpc_server(TRANSPORT,
|
||||
target,
|
||||
endpoints,
|
||||
executor='eventlet',
|
||||
serializer=serializer,
|
||||
access_policy=access_policy)
|
||||
return messaging.get_rpc_server(
|
||||
TRANSPORT,
|
||||
target,
|
||||
endpoints,
|
||||
executor='eventlet',
|
||||
serializer=serializer,
|
||||
access_policy=access_policy
|
||||
)
|
||||
|
||||
|
||||
def get_notification_listener(targets, endpoints, serializer=None, pool=None):
|
||||
assert NOTIFICATION_TRANSPORT is not None
|
||||
serializer = RequestContextSerializer(serializer)
|
||||
return messaging.get_notification_listener(
|
||||
NOTIFICATION_TRANSPORT,
|
||||
targets,
|
||||
endpoints,
|
||||
allow_requeue=False,
|
||||
executor='eventlet',
|
||||
pool=pool,
|
||||
serializer=serializer
|
||||
)
|
||||
|
||||
|
||||
def get_notifier(publisher_id):
|
||||
|
||||
@@ -21,14 +21,12 @@ from oslo_concurrency import processutils
|
||||
from oslo_config import cfg
|
||||
from oslo_log import _options
|
||||
from oslo_log import log
|
||||
import oslo_messaging as om
|
||||
import oslo_messaging as messaging
|
||||
from oslo_reports import guru_meditation_report as gmr
|
||||
from oslo_reports import opts as gmr_opts
|
||||
from oslo_service import service
|
||||
from oslo_service import wsgi
|
||||
|
||||
from oslo_messaging.rpc import dispatcher
|
||||
|
||||
from watcher._i18n import _
|
||||
from watcher.api import app
|
||||
from watcher.common import config
|
||||
@@ -183,11 +181,6 @@ class Service(service.ServiceBase):
|
||||
]
|
||||
self.notification_endpoints = self.manager.notification_endpoints
|
||||
|
||||
self.serializer = rpc.RequestContextSerializer(
|
||||
base.WatcherObjectSerializer())
|
||||
|
||||
self._transport = None
|
||||
self._notification_transport = None
|
||||
self._conductor_client = None
|
||||
|
||||
self.conductor_topic_handler = None
|
||||
@@ -201,27 +194,17 @@ class Service(service.ServiceBase):
|
||||
self.notification_topics, self.notification_endpoints
|
||||
)
|
||||
|
||||
@property
|
||||
def transport(self):
|
||||
if self._transport is None:
|
||||
self._transport = om.get_rpc_transport(CONF)
|
||||
return self._transport
|
||||
|
||||
@property
|
||||
def notification_transport(self):
|
||||
if self._notification_transport is None:
|
||||
self._notification_transport = om.get_notification_transport(CONF)
|
||||
return self._notification_transport
|
||||
|
||||
@property
|
||||
def conductor_client(self):
|
||||
if self._conductor_client is None:
|
||||
target = om.Target(
|
||||
target = messaging.Target(
|
||||
topic=self.conductor_topic,
|
||||
version=self.API_VERSION,
|
||||
)
|
||||
self._conductor_client = om.RPCClient(
|
||||
self.transport, target, serializer=self.serializer)
|
||||
self._conductor_client = rpc.get_client(
|
||||
target,
|
||||
serializer=base.WatcherObjectSerializer()
|
||||
)
|
||||
return self._conductor_client
|
||||
|
||||
@conductor_client.setter
|
||||
@@ -229,21 +212,18 @@ class Service(service.ServiceBase):
|
||||
self.conductor_client = c
|
||||
|
||||
def build_topic_handler(self, topic_name, endpoints=()):
|
||||
access_policy = dispatcher.DefaultRPCAccessPolicy
|
||||
serializer = rpc.RequestContextSerializer(rpc.JsonPayloadSerializer())
|
||||
target = om.Target(
|
||||
target = messaging.Target(
|
||||
topic=topic_name,
|
||||
# For compatibility, we can override it with 'host' opt
|
||||
server=CONF.host or socket.gethostname(),
|
||||
version=self.api_version,
|
||||
)
|
||||
return om.get_rpc_server(
|
||||
self.transport, target, endpoints,
|
||||
executor='eventlet', serializer=serializer,
|
||||
access_policy=access_policy)
|
||||
return rpc.get_server(
|
||||
target, endpoints,
|
||||
serializer=rpc.JsonPayloadSerializer()
|
||||
)
|
||||
|
||||
def build_notification_handler(self, topic_names, endpoints=()):
|
||||
serializer = rpc.RequestContextSerializer(rpc.JsonPayloadSerializer())
|
||||
targets = []
|
||||
for topic in topic_names:
|
||||
kwargs = {}
|
||||
@@ -251,11 +231,13 @@ class Service(service.ServiceBase):
|
||||
exchange, topic = topic.split('.')
|
||||
kwargs['exchange'] = exchange
|
||||
kwargs['topic'] = topic
|
||||
targets.append(om.Target(**kwargs))
|
||||
return om.get_notification_listener(
|
||||
self.notification_transport, targets, endpoints,
|
||||
executor='eventlet', serializer=serializer,
|
||||
allow_requeue=False, pool=CONF.host)
|
||||
targets.append(messaging.Target(**kwargs))
|
||||
|
||||
return rpc.get_notification_listener(
|
||||
targets, endpoints,
|
||||
serializer=rpc.JsonPayloadSerializer(),
|
||||
pool=CONF.host
|
||||
)
|
||||
|
||||
def start(self):
|
||||
LOG.debug("Connecting to '%s'", CONF.transport_url)
|
||||
|
||||
@@ -15,11 +15,9 @@
|
||||
# under the License.
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class ServiceManager(object):
|
||||
class ServiceManager(object, metaclass=abc.ABCMeta):
|
||||
|
||||
@abc.abstractproperty
|
||||
def service_name(self):
|
||||
|
||||
@@ -28,7 +28,6 @@ from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
from oslo_utils import strutils
|
||||
from oslo_utils import uuidutils
|
||||
import six
|
||||
|
||||
from watcher.common import exception
|
||||
|
||||
@@ -82,7 +81,7 @@ def safe_rstrip(value, chars=None):
|
||||
:return: Stripped value.
|
||||
|
||||
"""
|
||||
if not isinstance(value, six.string_types):
|
||||
if not isinstance(value, str):
|
||||
LOG.warning(
|
||||
"Failed to remove trailing character. Returning original object."
|
||||
"Supplied object is not a string: %s,", value)
|
||||
@@ -104,7 +103,7 @@ def is_hostname_safe(hostname):
|
||||
|
||||
"""
|
||||
m = r'^[a-z0-9]([a-z0-9\-]{0,61}[a-z0-9])?$'
|
||||
return (isinstance(hostname, six.string_types) and
|
||||
return (isinstance(hostname, str) and
|
||||
(re.match(m, hostname) is not None))
|
||||
|
||||
|
||||
|
||||
@@ -43,11 +43,20 @@ APPLIER_MANAGER_OPTS = [
|
||||
help='Select the engine to use to execute the workflow'),
|
||||
]
|
||||
|
||||
APPLIER_OPTS = [
|
||||
cfg.BoolOpt('rollback_when_actionplan_failed',
|
||||
default=False,
|
||||
help='If set True, the failed actionplan will rollback '
|
||||
'when executing. Defaule value is False.'),
|
||||
]
|
||||
|
||||
|
||||
def register_opts(conf):
|
||||
conf.register_group(watcher_applier)
|
||||
conf.register_opts(APPLIER_MANAGER_OPTS, group=watcher_applier)
|
||||
conf.register_opts(APPLIER_OPTS, group=watcher_applier)
|
||||
|
||||
|
||||
def list_opts():
|
||||
return [(watcher_applier, APPLIER_MANAGER_OPTS)]
|
||||
return [(watcher_applier, APPLIER_MANAGER_OPTS),
|
||||
(watcher_applier, APPLIER_OPTS)]
|
||||
|
||||
@@ -18,7 +18,6 @@ Base classes for storage engines
|
||||
import abc
|
||||
from oslo_config import cfg
|
||||
from oslo_db import api as db_api
|
||||
import six
|
||||
|
||||
_BACKEND_MAPPING = {'sqlalchemy': 'watcher.db.sqlalchemy.api'}
|
||||
IMPL = db_api.DBAPI.from_config(cfg.CONF, backend_mapping=_BACKEND_MAPPING,
|
||||
@@ -30,8 +29,7 @@ def get_instance():
|
||||
return IMPL
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseConnection(object):
|
||||
class BaseConnection(object, metaclass=abc.ABCMeta):
|
||||
"""Base class for storage system connections."""
|
||||
|
||||
@abc.abstractmethod
|
||||
|
||||
@@ -15,8 +15,6 @@
|
||||
# limitations under the License.
|
||||
#
|
||||
|
||||
from __future__ import print_function
|
||||
|
||||
import collections
|
||||
import datetime
|
||||
import itertools
|
||||
@@ -25,7 +23,6 @@ import sys
|
||||
from oslo_log import log
|
||||
from oslo_utils import strutils
|
||||
import prettytable as ptable
|
||||
from six.moves import input
|
||||
|
||||
from watcher._i18n import _
|
||||
from watcher._i18n import lazy_translation_enabled
|
||||
|
||||
@@ -278,7 +278,7 @@ class Connection(api.BaseConnection):
|
||||
query = model_query(model, session=session)
|
||||
query = add_identity_filter(query, id_)
|
||||
try:
|
||||
ref = query.with_lockmode('update').one()
|
||||
ref = query.with_for_update().one()
|
||||
except exc.NoResultFound:
|
||||
raise exception.ResourceNotFound(name=model.__name__, id=id_)
|
||||
|
||||
@@ -815,7 +815,7 @@ class Connection(api.BaseConnection):
|
||||
query = model_query(models.Action, session=session)
|
||||
query = add_identity_filter(query, action_id)
|
||||
try:
|
||||
ref = query.with_lockmode('update').one()
|
||||
ref = query.with_for_update().one()
|
||||
except exc.NoResultFound:
|
||||
raise exception.ActionNotFound(action=action_id)
|
||||
|
||||
@@ -900,7 +900,7 @@ class Connection(api.BaseConnection):
|
||||
query = model_query(models.ActionPlan, session=session)
|
||||
query = add_identity_filter(query, action_plan_id)
|
||||
try:
|
||||
ref = query.with_lockmode('update').one()
|
||||
ref = query.with_for_update().one()
|
||||
except exc.NoResultFound:
|
||||
raise exception.ActionPlanNotFound(action_plan=action_plan_id)
|
||||
|
||||
|
||||
@@ -18,7 +18,6 @@ SQLAlchemy models for watcher service
|
||||
|
||||
from oslo_db.sqlalchemy import models
|
||||
from oslo_serialization import jsonutils
|
||||
import six.moves.urllib.parse as urlparse
|
||||
from sqlalchemy import Boolean
|
||||
from sqlalchemy import Column
|
||||
from sqlalchemy import DateTime
|
||||
@@ -33,7 +32,7 @@ from sqlalchemy import String
|
||||
from sqlalchemy import Text
|
||||
from sqlalchemy.types import TypeDecorator, TEXT
|
||||
from sqlalchemy import UniqueConstraint
|
||||
|
||||
import urllib.parse as urlparse
|
||||
from watcher import conf
|
||||
|
||||
CONF = conf.CONF
|
||||
|
||||
@@ -18,7 +18,6 @@
|
||||
# limitations under the License.
|
||||
#
|
||||
import abc
|
||||
import six
|
||||
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
@@ -36,9 +35,11 @@ CONF = cfg.CONF
|
||||
LOG = log.getLogger(__name__)
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
@six.add_metaclass(service.Singleton)
|
||||
class BaseAuditHandler(object):
|
||||
class BaseMetaClass(service.Singleton, abc.ABCMeta):
|
||||
pass
|
||||
|
||||
|
||||
class BaseAuditHandler(object, metaclass=BaseMetaClass):
|
||||
|
||||
@abc.abstractmethod
|
||||
def execute(self, audit, request_context):
|
||||
@@ -57,8 +58,7 @@ class BaseAuditHandler(object):
|
||||
raise NotImplementedError()
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class AuditHandler(BaseAuditHandler):
|
||||
class AuditHandler(BaseAuditHandler, metaclass=abc.ABCMeta):
|
||||
|
||||
def __init__(self):
|
||||
super(AuditHandler, self).__init__()
|
||||
|
||||
@@ -19,6 +19,8 @@ import time
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
|
||||
from watcher.common import exception
|
||||
|
||||
CONF = cfg.CONF
|
||||
LOG = log.getLogger(__name__)
|
||||
|
||||
@@ -54,6 +56,13 @@ class DataSourceBase(object):
|
||||
instance_root_disk_size=None,
|
||||
)
|
||||
|
||||
def _get_meter(self, meter_name):
|
||||
"""Retrieve the meter from the metric map or raise error"""
|
||||
meter = self.METRIC_MAP.get(meter_name)
|
||||
if meter is None:
|
||||
raise exception.MetricNotAvailable(metric=meter_name)
|
||||
return meter
|
||||
|
||||
def query_retry(self, f, *args, **kwargs):
|
||||
"""Attempts to retrieve metrics from the external service
|
||||
|
||||
@@ -122,6 +131,30 @@ class DataSourceBase(object):
|
||||
|
||||
pass
|
||||
|
||||
@abc.abstractmethod
|
||||
def statistic_series(self, resource=None, resource_type=None,
|
||||
meter_name=None, start_time=None, end_time=None,
|
||||
granularity=300):
|
||||
"""Retrieves metrics based on the specified parameters over a period
|
||||
|
||||
:param resource: Resource object as defined in watcher models such as
|
||||
ComputeNode and Instance
|
||||
:param resource_type: Indicates which type of object is supplied
|
||||
to the resource parameter
|
||||
:param meter_name: The desired metric to retrieve as key from
|
||||
METRIC_MAP
|
||||
:param start_time: The datetime to start retrieving metrics for
|
||||
:type start_time: datetime.datetime
|
||||
:param end_time: The datetime to limit the retrieval of metrics to
|
||||
:type end_time: datetime.datetime
|
||||
:param granularity: Interval between samples in measurements in
|
||||
seconds
|
||||
:return: Dictionary of key value pairs with timestamps and metric
|
||||
values
|
||||
"""
|
||||
|
||||
pass
|
||||
|
||||
@abc.abstractmethod
|
||||
def get_host_cpu_usage(self, resource, period, aggregate,
|
||||
granularity=None):
|
||||
|
||||
@@ -161,9 +161,7 @@ class CeilometerHelper(base.DataSourceBase):
|
||||
end_time = datetime.datetime.utcnow()
|
||||
start_time = end_time - datetime.timedelta(seconds=int(period))
|
||||
|
||||
meter = self.METRIC_MAP.get(meter_name)
|
||||
if meter is None:
|
||||
raise exception.MetricNotAvailable(metric=meter_name)
|
||||
meter = self._get_meter(meter_name)
|
||||
|
||||
if aggregate == 'mean':
|
||||
aggregate = 'avg'
|
||||
@@ -194,6 +192,12 @@ class CeilometerHelper(base.DataSourceBase):
|
||||
item_value *= 10
|
||||
return item_value
|
||||
|
||||
def statistic_series(self, resource=None, resource_type=None,
|
||||
meter_name=None, start_time=None, end_time=None,
|
||||
granularity=300):
|
||||
raise NotImplementedError(
|
||||
_('Ceilometer helper does not support statistic series method'))
|
||||
|
||||
def get_host_cpu_usage(self, resource, period,
|
||||
aggregate, granularity=None):
|
||||
|
||||
|
||||
@@ -23,7 +23,6 @@ from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
|
||||
from watcher.common import clients
|
||||
from watcher.common import exception
|
||||
from watcher.decision_engine.datasources import base
|
||||
|
||||
CONF = cfg.CONF
|
||||
@@ -72,9 +71,7 @@ class GnocchiHelper(base.DataSourceBase):
|
||||
stop_time = datetime.utcnow()
|
||||
start_time = stop_time - timedelta(seconds=(int(period)))
|
||||
|
||||
meter = self.METRIC_MAP.get(meter_name)
|
||||
if meter is None:
|
||||
raise exception.MetricNotAvailable(metric=meter_name)
|
||||
meter = self._get_meter(meter_name)
|
||||
|
||||
if aggregate == 'count':
|
||||
aggregate = 'mean'
|
||||
@@ -123,6 +120,52 @@ class GnocchiHelper(base.DataSourceBase):
|
||||
|
||||
return return_value
|
||||
|
||||
def statistic_series(self, resource=None, resource_type=None,
|
||||
meter_name=None, start_time=None, end_time=None,
|
||||
granularity=300):
|
||||
|
||||
meter = self._get_meter(meter_name)
|
||||
|
||||
resource_id = resource.uuid
|
||||
if resource_type == 'compute_node':
|
||||
resource_id = "%s_%s" % (resource.hostname, resource.hostname)
|
||||
kwargs = dict(query={"=": {"original_resource_id": resource_id}},
|
||||
limit=1)
|
||||
resources = self.query_retry(
|
||||
f=self.gnocchi.resource.search, **kwargs)
|
||||
|
||||
if not resources:
|
||||
LOG.warning("The {0} resource {1} could not be "
|
||||
"found".format(self.NAME, resource_id))
|
||||
return
|
||||
|
||||
resource_id = resources[0]['id']
|
||||
|
||||
raw_kwargs = dict(
|
||||
metric=meter,
|
||||
start=start_time,
|
||||
stop=end_time,
|
||||
resource_id=resource_id,
|
||||
granularity=granularity,
|
||||
)
|
||||
|
||||
kwargs = {k: v for k, v in raw_kwargs.items() if k and v}
|
||||
|
||||
statistics = self.query_retry(
|
||||
f=self.gnocchi.metric.get_measures, **kwargs)
|
||||
|
||||
return_value = None
|
||||
if statistics:
|
||||
# measure has structure [time, granularity, value]
|
||||
if meter_name == 'host_airflow':
|
||||
# Airflow from hardware.ipmi.node.airflow is reported as
|
||||
# 1/10 th of actual CFM
|
||||
return_value = {s[0]: s[2]*10 for s in statistics}
|
||||
else:
|
||||
return_value = {s[0]: s[2] for s in statistics}
|
||||
|
||||
return return_value
|
||||
|
||||
def get_host_cpu_usage(self, resource, period, aggregate,
|
||||
granularity=300):
|
||||
|
||||
|
||||
@@ -16,10 +16,12 @@
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
from urllib import parse as urlparse
|
||||
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
import six.moves.urllib.parse as urlparse
|
||||
|
||||
from watcher._i18n import _
|
||||
from watcher.common import clients
|
||||
from watcher.common import exception
|
||||
from watcher.decision_engine.datasources import base
|
||||
@@ -187,6 +189,12 @@ class GrafanaHelper(base.DataSourceBase):
|
||||
|
||||
return result
|
||||
|
||||
def statistic_series(self, resource=None, resource_type=None,
|
||||
meter_name=None, start_time=None, end_time=None,
|
||||
granularity=300):
|
||||
raise NotImplementedError(
|
||||
_('Grafana helper does not support statistic series method'))
|
||||
|
||||
def get_host_cpu_usage(self, resource, period=300,
|
||||
aggregate="mean", granularity=None):
|
||||
return self.statistic_aggregation(
|
||||
|
||||
@@ -21,7 +21,6 @@ import datetime
|
||||
from monascaclient import exc
|
||||
|
||||
from watcher.common import clients
|
||||
from watcher.common import exception
|
||||
from watcher.decision_engine.datasources import base
|
||||
|
||||
|
||||
@@ -90,9 +89,7 @@ class MonascaHelper(base.DataSourceBase):
|
||||
stop_time = datetime.datetime.utcnow()
|
||||
start_time = stop_time - datetime.timedelta(seconds=(int(period)))
|
||||
|
||||
meter = self.METRIC_MAP.get(meter_name)
|
||||
if meter is None:
|
||||
raise exception.MetricNotAvailable(metric=meter_name)
|
||||
meter = self._get_meter(meter_name)
|
||||
|
||||
if aggregate == 'mean':
|
||||
aggregate = 'avg'
|
||||
@@ -121,6 +118,34 @@ class MonascaHelper(base.DataSourceBase):
|
||||
|
||||
return cpu_usage
|
||||
|
||||
def statistic_series(self, resource=None, resource_type=None,
|
||||
meter_name=None, start_time=None, end_time=None,
|
||||
granularity=300):
|
||||
|
||||
meter = self._get_meter(meter_name)
|
||||
|
||||
raw_kwargs = dict(
|
||||
name=meter,
|
||||
start_time=start_time.isoformat(),
|
||||
end_time=end_time.isoformat(),
|
||||
dimensions={'hostname': resource.uuid},
|
||||
statistics='avg',
|
||||
group_by='*',
|
||||
)
|
||||
|
||||
kwargs = {k: v for k, v in raw_kwargs.items() if k and v}
|
||||
|
||||
statistics = self.query_retry(
|
||||
f=self.monasca.metrics.list_statistics, **kwargs)
|
||||
|
||||
result = {}
|
||||
for stat in statistics:
|
||||
v_index = stat['columns'].index('avg')
|
||||
t_index = stat['columns'].index('timestamp')
|
||||
result.update({r[t_index]: r[v_index] for r in stat['statistics']})
|
||||
|
||||
return result
|
||||
|
||||
def get_host_cpu_usage(self, resource, period,
|
||||
aggregate, granularity=None):
|
||||
return self.statistic_aggregation(
|
||||
|
||||
@@ -15,13 +15,11 @@
|
||||
# limitations under the License.
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
from watcher.common.loader import loadable
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class Goal(loadable.Loadable):
|
||||
class Goal(loadable.Loadable, metaclass=abc.ABCMeta):
|
||||
|
||||
def __init__(self, config):
|
||||
super(Goal, self).__init__(config)
|
||||
|
||||
@@ -27,11 +27,8 @@ import abc
|
||||
import jsonschema
|
||||
from oslo_serialization import jsonutils
|
||||
|
||||
import six
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class EfficacySpecification(object):
|
||||
class EfficacySpecification(object, metaclass=abc.ABCMeta):
|
||||
|
||||
def __init__(self):
|
||||
self._indicators_specs = self.get_indicators_specifications()
|
||||
|
||||
@@ -18,7 +18,6 @@ import abc
|
||||
import jsonschema
|
||||
from jsonschema import SchemaError
|
||||
from jsonschema import ValidationError
|
||||
import six
|
||||
|
||||
from oslo_log import log
|
||||
from oslo_serialization import jsonutils
|
||||
@@ -29,8 +28,7 @@ from watcher.common import exception
|
||||
LOG = log.getLogger(__name__)
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class IndicatorSpecification(object):
|
||||
class IndicatorSpecification(object, metaclass=abc.ABCMeta):
|
||||
|
||||
def __init__(self, name=None, description=None, unit=None, required=True):
|
||||
self.name = name
|
||||
|
||||
@@ -19,9 +19,6 @@
|
||||
# limitations under the License.
|
||||
#
|
||||
|
||||
from __future__ import unicode_literals
|
||||
|
||||
|
||||
from watcher.common.loader import default
|
||||
|
||||
|
||||
|
||||
@@ -25,11 +25,9 @@ See: :doc:`../architecture` for more details on this component.
|
||||
"""
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class Model(object):
|
||||
class Model(object, metaclass=abc.ABCMeta):
|
||||
|
||||
@abc.abstractmethod
|
||||
def to_string(self):
|
||||
|
||||
@@ -110,7 +110,6 @@ import time
|
||||
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
import six
|
||||
|
||||
from watcher.common import clients
|
||||
from watcher.common.loader import loadable
|
||||
@@ -120,8 +119,8 @@ LOG = log.getLogger(__name__)
|
||||
CONF = cfg.CONF
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseClusterDataModelCollector(loadable.LoadableSingleton):
|
||||
class BaseClusterDataModelCollector(loadable.LoadableSingleton,
|
||||
metaclass=abc.ABCMeta):
|
||||
|
||||
STALE_MODEL = model_root.ModelRoot(stale=True)
|
||||
|
||||
|
||||
@@ -13,8 +13,6 @@
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
import six
|
||||
|
||||
from oslo_log import log
|
||||
|
||||
from watcher.common import cinder_helper
|
||||
@@ -152,6 +150,9 @@ class CinderClusterDataModelCollector(base.BaseClusterDataModelCollector):
|
||||
if self._audit_scope_handler is None:
|
||||
LOG.debug("No audit, Don't Build storage data model")
|
||||
return
|
||||
if self._data_model_scope is None:
|
||||
LOG.debug("No audit scope, Don't Build storage data model")
|
||||
return
|
||||
|
||||
builder = CinderModelBuilder(self.osc)
|
||||
return builder.execute(self._data_model_scope)
|
||||
@@ -286,7 +287,7 @@ class CinderModelBuilder(base.BaseModelBuilder):
|
||||
:param instance: Cinder Volume object.
|
||||
:return: A volume node for the graph.
|
||||
"""
|
||||
attachments = [{k: v for k, v in six.iteritems(d) if k in (
|
||||
attachments = [{k: v for k, v in iter(d.items()) if k in (
|
||||
'server_id', 'attachment_id')} for d in volume.attachments]
|
||||
|
||||
volume_attributes = {
|
||||
|
||||
@@ -63,6 +63,9 @@ class BaremetalClusterDataModelCollector(base.BaseClusterDataModelCollector):
|
||||
if self._audit_scope_handler is None:
|
||||
LOG.debug("No audit, Don't Build Baremetal data model")
|
||||
return
|
||||
if self._data_model_scope is None:
|
||||
LOG.debug("No audit scope, Don't Build Baremetal data model")
|
||||
return
|
||||
|
||||
builder = BareMetalModelBuilder(self.osc)
|
||||
return builder.execute(self._data_model_scope)
|
||||
|
||||
@@ -184,6 +184,9 @@ class NovaClusterDataModelCollector(base.BaseClusterDataModelCollector):
|
||||
if self._audit_scope_handler is None:
|
||||
LOG.debug("No audit, Don't Build compute data model")
|
||||
return
|
||||
if self._data_model_scope is None:
|
||||
LOG.debug("No audit scope, Don't Build compute data model")
|
||||
return
|
||||
|
||||
builder = NovaModelBuilder(self.osc)
|
||||
return builder.execute(self._data_model_scope)
|
||||
|
||||
@@ -16,14 +16,12 @@
|
||||
|
||||
import abc
|
||||
|
||||
import six
|
||||
|
||||
from watcher.decision_engine.model.element import base
|
||||
from watcher.objects import fields as wfields
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaremetalResource(base.Element):
|
||||
class BaremetalResource(base.Element, metaclass=abc.ABCMeta):
|
||||
|
||||
VERSION = '1.0'
|
||||
|
||||
|
||||
@@ -21,7 +21,6 @@ import collections
|
||||
|
||||
from lxml import etree
|
||||
from oslo_log import log
|
||||
import six
|
||||
|
||||
from watcher.objects import base
|
||||
from watcher.objects import fields as wfields
|
||||
@@ -29,9 +28,8 @@ from watcher.objects import fields as wfields
|
||||
LOG = log.getLogger(__name__)
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class Element(base.WatcherObject, base.WatcherObjectDictCompat,
|
||||
base.WatcherComparableObject):
|
||||
base.WatcherComparableObject, metaclass=abc.ABCMeta):
|
||||
|
||||
# Initial version
|
||||
VERSION = '1.0'
|
||||
|
||||
@@ -16,14 +16,12 @@
|
||||
|
||||
import abc
|
||||
|
||||
import six
|
||||
|
||||
from watcher.decision_engine.model.element import base
|
||||
from watcher.objects import fields as wfields
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class ComputeResource(base.Element):
|
||||
class ComputeResource(base.Element, metaclass=abc.ABCMeta):
|
||||
|
||||
VERSION = '1.0'
|
||||
|
||||
|
||||
@@ -16,14 +16,12 @@
|
||||
|
||||
import abc
|
||||
|
||||
import six
|
||||
|
||||
from watcher.decision_engine.model.element import base
|
||||
from watcher.objects import fields as wfields
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class StorageResource(base.Element):
|
||||
class StorageResource(base.Element, metaclass=abc.ABCMeta):
|
||||
|
||||
VERSION = '1.0'
|
||||
|
||||
|
||||
@@ -21,7 +21,6 @@ from lxml import etree
|
||||
import networkx as nx
|
||||
from oslo_concurrency import lockutils
|
||||
from oslo_log import log
|
||||
import six
|
||||
|
||||
from watcher._i18n import _
|
||||
from watcher.common import exception
|
||||
@@ -93,9 +92,9 @@ class ModelRoot(nx.DiGraph, base.Model):
|
||||
:param node: :py:class:`~.node.ComputeNode` object or node UUID
|
||||
:type node: str or :py:class:`~.instance.Instance`
|
||||
"""
|
||||
if isinstance(instance, six.string_types):
|
||||
if isinstance(instance, str):
|
||||
instance = self.get_instance_by_uuid(instance)
|
||||
if isinstance(node, six.string_types):
|
||||
if isinstance(node, str):
|
||||
node = self.get_node_by_uuid(node)
|
||||
self.assert_node(node)
|
||||
self.assert_instance(instance)
|
||||
@@ -104,9 +103,9 @@ class ModelRoot(nx.DiGraph, base.Model):
|
||||
|
||||
@lockutils.synchronized("model_root")
|
||||
def unmap_instance(self, instance, node):
|
||||
if isinstance(instance, six.string_types):
|
||||
if isinstance(instance, str):
|
||||
instance = self.get_instance_by_uuid(instance)
|
||||
if isinstance(node, six.string_types):
|
||||
if isinstance(node, str):
|
||||
node = self.get_node_by_uuid(node)
|
||||
|
||||
self.remove_edge(instance.uuid, node.uuid)
|
||||
@@ -367,9 +366,9 @@ class StorageModelRoot(nx.DiGraph, base.Model):
|
||||
:param pool: :py:class:`~.node.Pool` object or pool name
|
||||
:param node: :py:class:`~.node.StorageNode` object or node host
|
||||
"""
|
||||
if isinstance(pool, six.string_types):
|
||||
if isinstance(pool, str):
|
||||
pool = self.get_pool_by_pool_name(pool)
|
||||
if isinstance(node, six.string_types):
|
||||
if isinstance(node, str):
|
||||
node = self.get_node_by_name(node)
|
||||
self.assert_node(node)
|
||||
self.assert_pool(pool)
|
||||
@@ -383,9 +382,9 @@ class StorageModelRoot(nx.DiGraph, base.Model):
|
||||
:param pool: :py:class:`~.node.Pool` object or pool name
|
||||
:param node: :py:class:`~.node.StorageNode` object or node name
|
||||
"""
|
||||
if isinstance(pool, six.string_types):
|
||||
if isinstance(pool, str):
|
||||
pool = self.get_pool_by_pool_name(pool)
|
||||
if isinstance(node, six.string_types):
|
||||
if isinstance(node, str):
|
||||
node = self.get_node_by_name(node)
|
||||
|
||||
self.remove_edge(pool.name, node.host)
|
||||
@@ -411,9 +410,9 @@ class StorageModelRoot(nx.DiGraph, base.Model):
|
||||
:param volume: :py:class:`~.volume.Volume` object or volume UUID
|
||||
:param pool: :py:class:`~.node.Pool` object or pool name
|
||||
"""
|
||||
if isinstance(volume, six.string_types):
|
||||
if isinstance(volume, str):
|
||||
volume = self.get_volume_by_uuid(volume)
|
||||
if isinstance(pool, six.string_types):
|
||||
if isinstance(pool, str):
|
||||
pool = self.get_pool_by_pool_name(pool)
|
||||
self.assert_pool(pool)
|
||||
self.assert_volume(volume)
|
||||
@@ -427,9 +426,9 @@ class StorageModelRoot(nx.DiGraph, base.Model):
|
||||
:param volume: :py:class:`~.volume.Volume` object or volume UUID
|
||||
:param pool: :py:class:`~.node.Pool` object or pool name
|
||||
"""
|
||||
if isinstance(volume, six.string_types):
|
||||
if isinstance(volume, str):
|
||||
volume = self.get_volume_by_uuid(volume)
|
||||
if isinstance(pool, six.string_types):
|
||||
if isinstance(pool, str):
|
||||
pool = self.get_pool_by_pool_name(pool)
|
||||
|
||||
self.remove_edge(volume.uuid, pool.name)
|
||||
@@ -632,7 +631,7 @@ class BaremetalModelRoot(nx.DiGraph, base.Model):
|
||||
super(BaremetalModelRoot, self).remove_node(node.uuid)
|
||||
except nx.NetworkXError as exc:
|
||||
LOG.exception(exc)
|
||||
raise exception.IronicNodeNotFound(name=node.uuid)
|
||||
raise exception.IronicNodeNotFound(uuid=node.uuid)
|
||||
|
||||
@lockutils.synchronized("baremetal_model")
|
||||
def get_all_ironic_nodes(self):
|
||||
@@ -644,7 +643,7 @@ class BaremetalModelRoot(nx.DiGraph, base.Model):
|
||||
try:
|
||||
return self._get_by_uuid(uuid)
|
||||
except exception.BaremetalResourceNotFound:
|
||||
raise exception.IronicNodeNotFound(name=uuid)
|
||||
raise exception.IronicNodeNotFound(uuid=uuid)
|
||||
|
||||
def _get_by_uuid(self, uuid):
|
||||
try:
|
||||
|
||||
@@ -17,11 +17,9 @@
|
||||
# limitations under the License.
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class NotificationEndpoint(object):
|
||||
class NotificationEndpoint(object, metaclass=abc.ABCMeta):
|
||||
|
||||
def __init__(self, collector):
|
||||
super(NotificationEndpoint, self).__init__()
|
||||
|
||||
@@ -14,8 +14,6 @@
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
import six
|
||||
|
||||
from oslo_log import log
|
||||
from watcher.common import cinder_helper
|
||||
from watcher.common import exception
|
||||
@@ -161,7 +159,7 @@ class CinderNotification(base.NotificationEndpoint):
|
||||
return 'attachment_id'
|
||||
|
||||
attachments = [
|
||||
{_keyReplace(k): v for k, v in six.iteritems(d)
|
||||
{_keyReplace(k): v for k, v in iter(d.items())
|
||||
if k in ('instance_uuid', 'id')}
|
||||
for d in data['volume_attachment']
|
||||
]
|
||||
|
||||
@@ -19,7 +19,6 @@
|
||||
import re
|
||||
|
||||
import oslo_messaging as om
|
||||
import six
|
||||
|
||||
|
||||
class NotificationFilter(om.NotificationFilter):
|
||||
@@ -81,7 +80,7 @@ class NotificationFilter(om.NotificationFilter):
|
||||
elif regex is not None and data is None:
|
||||
return True
|
||||
elif (regex is not None and
|
||||
isinstance(data, six.string_types) and
|
||||
isinstance(data, str) and
|
||||
not regex.match(data)):
|
||||
return True
|
||||
|
||||
|
||||
@@ -45,13 +45,11 @@ See :doc:`../architecture` for more details on this component.
|
||||
"""
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
from watcher.common.loader import loadable
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BasePlanner(loadable.Loadable):
|
||||
class BasePlanner(loadable.Loadable, metaclass=abc.ABCMeta):
|
||||
|
||||
@classmethod
|
||||
def get_config_opts(cls):
|
||||
|
||||
@@ -16,13 +16,11 @@
|
||||
#
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
from watcher.common import context
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseScope(object):
|
||||
class BaseScope(object, metaclass=abc.ABCMeta):
|
||||
"""A base class for Scope mechanism
|
||||
|
||||
Child of this class is called when audit launches strategy. This strategy
|
||||
|
||||
@@ -17,13 +17,11 @@
|
||||
# limitations under the License.
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
from watcher.common.loader import loadable
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class ScoringEngine(loadable.Loadable):
|
||||
class ScoringEngine(loadable.Loadable, metaclass=abc.ABCMeta):
|
||||
"""A base class for all the Scoring Engines.
|
||||
|
||||
A Scoring Engine is an instance of a data model, to which the learning
|
||||
@@ -97,8 +95,7 @@ class ScoringEngine(loadable.Loadable):
|
||||
return []
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class ScoringEngineContainer(loadable.Loadable):
|
||||
class ScoringEngineContainer(loadable.Loadable, metaclass=abc.ABCMeta):
|
||||
"""A base class for all the Scoring Engines Containers.
|
||||
|
||||
A Scoring Engine Container is an abstraction which allows to plugin
|
||||
|
||||
@@ -56,13 +56,11 @@ Two approaches to dealing with this can be envisaged:
|
||||
"""
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
from watcher.decision_engine.solution import efficacy
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseSolution(object):
|
||||
class BaseSolution(object, metaclass=abc.ABCMeta):
|
||||
def __init__(self, goal, strategy):
|
||||
"""Base Solution constructor
|
||||
|
||||
|
||||
@@ -17,11 +17,9 @@
|
||||
# limitations under the License.
|
||||
#
|
||||
import abc
|
||||
import six
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseSolutionComparator(object):
|
||||
class BaseSolutionComparator(object, metaclass=abc.ABCMeta):
|
||||
@abc.abstractmethod
|
||||
def compare(self, sol1, sol2):
|
||||
raise NotImplementedError()
|
||||
|
||||
@@ -17,11 +17,9 @@
|
||||
# limitations under the License.
|
||||
#
|
||||
import abc
|
||||
import six
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseSolutionEvaluator(object):
|
||||
class BaseSolutionEvaluator(object, metaclass=abc.ABCMeta):
|
||||
@abc.abstractmethod
|
||||
def evaluate(self, solution):
|
||||
raise NotImplementedError()
|
||||
|
||||
@@ -18,14 +18,12 @@
|
||||
# limitations under the License.
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
from watcher import notifications
|
||||
from watcher.objects import fields
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class StrategyContext(object):
|
||||
class StrategyContext(object, metaclass=abc.ABCMeta):
|
||||
|
||||
def execute_strategy(self, audit, request_context):
|
||||
"""Execute the strategy for the given an audit
|
||||
|
||||
@@ -17,11 +17,9 @@
|
||||
# limitations under the License.
|
||||
#
|
||||
import abc
|
||||
import six
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseSelector(object):
|
||||
class BaseSelector(object, metaclass=abc.ABCMeta):
|
||||
|
||||
@abc.abstractmethod
|
||||
def select(self):
|
||||
|
||||
@@ -37,7 +37,6 @@ which are dynamically loaded by Watcher at launch time.
|
||||
"""
|
||||
|
||||
import abc
|
||||
import six
|
||||
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
@@ -121,8 +120,7 @@ class StrategyEndpoint(object):
|
||||
return [available_datasource, available_metrics, available_cdm]
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class BaseStrategy(loadable.Loadable):
|
||||
class BaseStrategy(loadable.Loadable, metaclass=abc.ABCMeta):
|
||||
"""A base class for all the strategies
|
||||
|
||||
A Strategy is an algorithm implementation which is able to find a
|
||||
@@ -471,8 +469,7 @@ class BaseStrategy(loadable.Loadable):
|
||||
input_parameters=parameters)
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class DummyBaseStrategy(BaseStrategy):
|
||||
class DummyBaseStrategy(BaseStrategy, metaclass=abc.ABCMeta):
|
||||
|
||||
@classmethod
|
||||
def get_goal_name(cls):
|
||||
@@ -485,8 +482,7 @@ class DummyBaseStrategy(BaseStrategy):
|
||||
return []
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class UnclassifiedStrategy(BaseStrategy):
|
||||
class UnclassifiedStrategy(BaseStrategy, metaclass=abc.ABCMeta):
|
||||
"""This base class is used to ease the development of new strategies
|
||||
|
||||
The goal defined within this strategy can be used to simplify the
|
||||
@@ -500,8 +496,7 @@ class UnclassifiedStrategy(BaseStrategy):
|
||||
return "unclassified"
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class ServerConsolidationBaseStrategy(BaseStrategy):
|
||||
class ServerConsolidationBaseStrategy(BaseStrategy, metaclass=abc.ABCMeta):
|
||||
|
||||
REASON_FOR_DISABLE = 'watcher_disabled'
|
||||
|
||||
@@ -510,16 +505,14 @@ class ServerConsolidationBaseStrategy(BaseStrategy):
|
||||
return "server_consolidation"
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class ThermalOptimizationBaseStrategy(BaseStrategy):
|
||||
class ThermalOptimizationBaseStrategy(BaseStrategy, metaclass=abc.ABCMeta):
|
||||
|
||||
@classmethod
|
||||
def get_goal_name(cls):
|
||||
return "thermal_optimization"
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class WorkloadStabilizationBaseStrategy(BaseStrategy):
|
||||
class WorkloadStabilizationBaseStrategy(BaseStrategy, metaclass=abc.ABCMeta):
|
||||
|
||||
def __init__(self, *args, **kwargs):
|
||||
super(WorkloadStabilizationBaseStrategy, self
|
||||
@@ -531,16 +524,14 @@ class WorkloadStabilizationBaseStrategy(BaseStrategy):
|
||||
return "workload_balancing"
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class NoisyNeighborBaseStrategy(BaseStrategy):
|
||||
class NoisyNeighborBaseStrategy(BaseStrategy, metaclass=abc.ABCMeta):
|
||||
|
||||
@classmethod
|
||||
def get_goal_name(cls):
|
||||
return "noisy_neighbor"
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class SavingEnergyBaseStrategy(BaseStrategy):
|
||||
class SavingEnergyBaseStrategy(BaseStrategy, metaclass=abc.ABCMeta):
|
||||
|
||||
@classmethod
|
||||
def get_goal_name(cls):
|
||||
@@ -553,8 +544,7 @@ class SavingEnergyBaseStrategy(BaseStrategy):
|
||||
return []
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class ZoneMigrationBaseStrategy(BaseStrategy):
|
||||
class ZoneMigrationBaseStrategy(BaseStrategy, metaclass=abc.ABCMeta):
|
||||
|
||||
@classmethod
|
||||
def get_goal_name(cls):
|
||||
@@ -567,8 +557,7 @@ class ZoneMigrationBaseStrategy(BaseStrategy):
|
||||
return []
|
||||
|
||||
|
||||
@six.add_metaclass(abc.ABCMeta)
|
||||
class HostMaintenanceBaseStrategy(BaseStrategy):
|
||||
class HostMaintenanceBaseStrategy(BaseStrategy, metaclass=abc.ABCMeta):
|
||||
|
||||
REASON_FOR_MAINTAINING = 'watcher_maintaining'
|
||||
|
||||
|
||||
@@ -18,8 +18,6 @@
|
||||
#
|
||||
|
||||
from oslo_log import log
|
||||
import six
|
||||
|
||||
from watcher._i18n import _
|
||||
from watcher.common import exception
|
||||
from watcher.decision_engine.model import element
|
||||
@@ -103,7 +101,7 @@ class HostMaintenance(base.HostMaintenanceBaseStrategy):
|
||||
|
||||
def get_instance_state_str(self, instance):
|
||||
"""Get instance state in string format"""
|
||||
if isinstance(instance.state, six.string_types):
|
||||
if isinstance(instance.state, str):
|
||||
return instance.state
|
||||
elif isinstance(instance.state, element.InstanceState):
|
||||
return instance.state.value
|
||||
@@ -116,7 +114,7 @@ class HostMaintenance(base.HostMaintenanceBaseStrategy):
|
||||
|
||||
def get_node_status_str(self, node):
|
||||
"""Get node status in string format"""
|
||||
if isinstance(node.status, six.string_types):
|
||||
if isinstance(node.status, str):
|
||||
return node.status
|
||||
elif isinstance(node.status, element.ServiceState):
|
||||
return node.status.value
|
||||
|
||||
@@ -19,7 +19,6 @@
|
||||
#
|
||||
|
||||
from oslo_log import log
|
||||
import six
|
||||
|
||||
from watcher._i18n import _
|
||||
from watcher.common import exception
|
||||
@@ -132,7 +131,7 @@ class VMWorkloadConsolidation(base.ServerConsolidationBaseStrategy):
|
||||
|
||||
:param instance:
|
||||
"""
|
||||
if isinstance(instance.state, six.string_types):
|
||||
if isinstance(instance.state, str):
|
||||
return instance.state
|
||||
elif isinstance(instance.state, element.InstanceState):
|
||||
return instance.state.value
|
||||
@@ -148,7 +147,7 @@ class VMWorkloadConsolidation(base.ServerConsolidationBaseStrategy):
|
||||
|
||||
:param node:
|
||||
"""
|
||||
if isinstance(node.status, six.string_types):
|
||||
if isinstance(node.status, str):
|
||||
return node.status
|
||||
elif isinstance(node.status, element.ServiceState):
|
||||
return node.status.value
|
||||
|
||||
@@ -17,8 +17,6 @@
|
||||
# limitations under the License.
|
||||
#
|
||||
|
||||
from __future__ import division
|
||||
|
||||
from oslo_log import log
|
||||
|
||||
from watcher._i18n import _
|
||||
|
||||
@@ -13,7 +13,6 @@
|
||||
#
|
||||
|
||||
from dateutil.parser import parse
|
||||
import six
|
||||
|
||||
from oslo_log import log
|
||||
|
||||
@@ -282,7 +281,7 @@ class ZoneMigration(base.ZoneMigrationBaseStrategy):
|
||||
action_counter = ActionCounter(total_limit,
|
||||
per_pool_limit, per_node_limit)
|
||||
|
||||
for k, targets in six.iteritems(filtered_targets):
|
||||
for k, targets in iter(filtered_targets.items()):
|
||||
if k == VOLUME:
|
||||
self.volumes_migration(targets, action_counter)
|
||||
elif k == INSTANCE:
|
||||
@@ -351,7 +350,7 @@ class ZoneMigration(base.ZoneMigrationBaseStrategy):
|
||||
def is_in_use(self, volume):
|
||||
return getattr(volume, 'status') == IN_USE
|
||||
|
||||
def instances_no_attached(instances):
|
||||
def instances_no_attached(self, instances):
|
||||
return [i for i in instances
|
||||
if not getattr(i, "os-extended-volumes:volumes_attached")]
|
||||
|
||||
@@ -580,7 +579,7 @@ class ZoneMigration(base.ZoneMigrationBaseStrategy):
|
||||
filter_list = []
|
||||
priority_filter_map = self.get_priority_filter_map()
|
||||
|
||||
for k, v in six.iteritems(self.priority):
|
||||
for k, v in iter(self.priority.items()):
|
||||
if k in priority_filter_map:
|
||||
filter_list.append(priority_filter_map[k](v))
|
||||
|
||||
@@ -710,7 +709,7 @@ class BaseFilter(object):
|
||||
return {}
|
||||
|
||||
for cond in list(reversed(self.condition)):
|
||||
for k, v in six.iteritems(targets):
|
||||
for k, v in iter(targets.items()):
|
||||
if not self.is_allowed(k):
|
||||
continue
|
||||
LOG.debug("filter:%s with the key: %s", cond, k)
|
||||
|
||||
@@ -19,7 +19,6 @@
|
||||
import copy
|
||||
import futurist
|
||||
from futurist import waiters
|
||||
import six
|
||||
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
@@ -29,8 +28,7 @@ CONF = cfg.CONF
|
||||
LOG = log.getLogger(__name__)
|
||||
|
||||
|
||||
@six.add_metaclass(service.Singleton)
|
||||
class DecisionEngineThreadPool(object):
|
||||
class DecisionEngineThreadPool(object, metaclass=service.Singleton):
|
||||
"""Singleton threadpool to submit general tasks to"""
|
||||
|
||||
def __init__(self):
|
||||
|
||||
@@ -128,22 +128,20 @@ def check_assert_called_once_with(logical_line, filename):
|
||||
@flake8ext
|
||||
def check_python3_xrange(logical_line):
|
||||
if re.search(r"\bxrange\s*\(", logical_line):
|
||||
yield(0, "N325: Do not use xrange. Use range, or six.moves.range for "
|
||||
"large loops.")
|
||||
yield(0, "N325: Do not use xrange. Use range for large loops.")
|
||||
|
||||
|
||||
@flake8ext
|
||||
def check_no_basestring(logical_line):
|
||||
if re.search(r"\bbasestring\b", logical_line):
|
||||
msg = ("N326: basestring is not Python3-compatible, use "
|
||||
"six.string_types instead.")
|
||||
msg = ("N326: basestring is not Python3-compatible, use str instead.")
|
||||
yield(0, msg)
|
||||
|
||||
|
||||
@flake8ext
|
||||
def check_python3_no_iteritems(logical_line):
|
||||
if re.search(r".*\.iteritems\(\)", logical_line):
|
||||
msg = ("N327: Use six.iteritems() instead of dict.iteritems().")
|
||||
msg = ("N327: Use dict.items() instead of dict.iteritems().")
|
||||
yield(0, msg)
|
||||
|
||||
|
||||
@@ -285,3 +283,31 @@ def no_redundant_import_alias(logical_line):
|
||||
"""
|
||||
if re.match(re_redundant_import_alias, logical_line):
|
||||
yield(0, "N342: No redundant import alias.")
|
||||
|
||||
|
||||
@flake8ext
|
||||
def import_stock_mock(logical_line):
|
||||
"""Use python's mock, not the mock library.
|
||||
|
||||
Since we `dropped support for python 2`__, we no longer need to use the
|
||||
mock library, which existed to backport py3 functionality into py2.
|
||||
Which must be done by saying::
|
||||
|
||||
from unittest import mock
|
||||
|
||||
...because if you say::
|
||||
|
||||
import mock
|
||||
|
||||
...you definitely will not be getting the standard library mock. That will
|
||||
always import the third party mock library. This check can be removed in
|
||||
the future (and we can start saying ``import mock`` again) if we manage to
|
||||
purge these transitive dependencies.
|
||||
|
||||
.. __: https://review.opendev.org/#/c/717540
|
||||
|
||||
N366
|
||||
"""
|
||||
if logical_line == 'import mock':
|
||||
yield (0, "N366: You must explicitly import python's mock: "
|
||||
"``from unittest import mock``")
|
||||
|
||||
@@ -1,14 +1,15 @@
|
||||
# Frank Kloeker <eumel@arcor.de>, 2018. #zanata
|
||||
# Andreas Jaeger <jaegerandi@gmail.com>, 2020. #zanata
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: watcher VERSION\n"
|
||||
"Report-Msgid-Bugs-To: https://bugs.launchpad.net/openstack-i18n/\n"
|
||||
"POT-Creation-Date: 2019-03-22 02:21+0000\n"
|
||||
"POT-Creation-Date: 2020-04-26 02:09+0000\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=UTF-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"PO-Revision-Date: 2018-03-06 07:56+0000\n"
|
||||
"Last-Translator: Frank Kloeker <eumel@arcor.de>\n"
|
||||
"PO-Revision-Date: 2020-04-25 11:45+0000\n"
|
||||
"Last-Translator: Andreas Jaeger <jaegerandi@gmail.com>\n"
|
||||
"Language-Team: German\n"
|
||||
"Language: de\n"
|
||||
"X-Generator: Zanata 4.3.3\n"
|
||||
@@ -62,10 +63,6 @@ msgstr "Der Argumenttyp 'obj' ist nicht gültig"
|
||||
msgid "'obj' argument type is not valid: %s"
|
||||
msgstr "Der Argumenttyp 'obj' ist nicht gültig: %s"
|
||||
|
||||
#, python-format
|
||||
msgid "A datetime.datetime is required here. Got %s"
|
||||
msgstr "Eine datetime.datetime ist hier erforderlich. Bekam %s"
|
||||
|
||||
#, python-format
|
||||
msgid "A goal with UUID %(uuid)s already exists"
|
||||
msgstr "Ein Ziel mit UUID %(uuid)s ist bereits vorhanden"
|
||||
@@ -169,10 +166,6 @@ msgstr ""
|
||||
msgid "An indicator value should be a number"
|
||||
msgstr "Ein Indikatorwert sollte eine Zahl sein"
|
||||
|
||||
#, python-format
|
||||
msgid "An object of class %s is required here"
|
||||
msgstr "Ein Objekt der Klasse %s ist hier erforderlich"
|
||||
|
||||
msgid "An unknown exception occurred"
|
||||
msgstr "Eine unbekannte Ausnahme ist aufgetreten"
|
||||
|
||||
@@ -203,20 +196,21 @@ msgstr "Prüfungsvorlagen"
|
||||
msgid "Audit parameter %(parameter)s are not allowed"
|
||||
msgstr "Prüfparameter %(parameter)s sind nicht erlaubt"
|
||||
|
||||
#, python-format
|
||||
msgid "Audit state %(state)s is disallowed."
|
||||
msgstr "Auditstatus %(state)s ist ungültig."
|
||||
|
||||
#, python-format
|
||||
msgid "Audit type %(audit_type)s could not be found"
|
||||
msgstr "Audit-Typ %(audit_type)s konnte nicht gefunden werden"
|
||||
|
||||
#, python-format
|
||||
msgid "AuditTemplate %(audit_template)s could not be found"
|
||||
msgstr "AuditTemplate %(audit_template)s konnte nicht gefunden werden"
|
||||
msgid "Audit type %(audit_type)s is disallowed."
|
||||
msgstr "Audit-Typ %(audit_type)s nicht erlaubt."
|
||||
|
||||
#, python-format
|
||||
msgid ""
|
||||
"AuditTemplate %(audit_template)s is referenced by one or multiple audits"
|
||||
msgstr ""
|
||||
"AuditTemplate %(audit_template)s wird von einem oder mehreren Audits "
|
||||
"referenziert"
|
||||
msgid "AuditTemplate %(audit_template)s could not be found"
|
||||
msgstr "AuditTemplate %(audit_template)s konnte nicht gefunden werden"
|
||||
|
||||
msgid "Audits"
|
||||
msgstr "Audits"
|
||||
@@ -293,6 +287,10 @@ msgstr "Patch '%(patch)s' konnte nicht angewendet werden. Grund:%(reason)s"
|
||||
msgid "Couldn't delete when state is '%(state)s'."
|
||||
msgstr "Konnte nicht gelöscht werden, wenn der Status '%(state)s' ist."
|
||||
|
||||
#, python-format
|
||||
msgid "Couldn't start when state is '%(state)s'."
|
||||
msgstr "Konnte nicht gestartet werden, wenn der Status '%(state)s' ist."
|
||||
|
||||
#, python-format
|
||||
msgid "Datasource %(datasource)s is not available."
|
||||
msgstr "Datenquelle %(datasource)s ist nicht verfügbar."
|
||||
@@ -435,6 +433,10 @@ msgstr "Ungültige Abfrage: %(start_time)s > %(end_time)s"
|
||||
msgid "Invalid sort direction: %s. Acceptable values are 'asc' or 'desc'"
|
||||
msgstr "Ungültige Sortierrichtung: %s. Akzeptable Werte sind 'asc' oder 'desc'"
|
||||
|
||||
#, python-format
|
||||
msgid "Invalid sort key: %s"
|
||||
msgstr "Ungültiger Sortierschlüssel: %s"
|
||||
|
||||
msgid "Invalid state for swapping volume"
|
||||
msgstr "Ungültiger Status für das Auslagern des Datenträgers"
|
||||
|
||||
@@ -445,9 +447,6 @@ msgstr "Ungültiger Status: %(state)s"
|
||||
msgid "JSON list expected in feature argument"
|
||||
msgstr "JSON-Liste in Feature-Argument erwartet"
|
||||
|
||||
msgid "Keystone API endpoint is missing"
|
||||
msgstr "Der Keystone-API-Endpunkt fehlt"
|
||||
|
||||
msgid "Limit must be positive"
|
||||
msgstr "Limit muss positiv sein"
|
||||
|
||||
@@ -457,30 +456,27 @@ msgstr "Limit sollte positiv sein"
|
||||
msgid "Maximum time since last check-in for up service."
|
||||
msgstr "Maximale Zeit seit dem letzten Check-in für den Up-Service."
|
||||
|
||||
#, python-format
|
||||
msgid "Metric: %(metric)s not available"
|
||||
msgstr "Metrik: %(metric)s nicht verfügbar"
|
||||
|
||||
#, python-format
|
||||
msgid "Migration of type '%(migration_type)s' is not supported."
|
||||
msgstr "Die Migration vom Typ '%(migration_type)s' wird nicht unterstützt."
|
||||
|
||||
msgid "Minimum Nova API Version"
|
||||
msgstr "Minimale Nova API Version"
|
||||
|
||||
#, python-format
|
||||
msgid "No %(metric)s metric for %(host)s found."
|
||||
msgstr "Keine %(metric)s Metrik für %(host)s gefunden."
|
||||
|
||||
msgid "No rows were returned"
|
||||
msgstr "Es wurden keine Zeilen zurückgegeben"
|
||||
|
||||
#, python-format
|
||||
msgid "No strategy could be found to achieve the '%(goal)s' goal."
|
||||
msgstr ""
|
||||
"Es konnte keine Strategie gefunden werden, um das Ziel '%(goal)s' zu "
|
||||
"erreichen."
|
||||
|
||||
msgid "No such metric"
|
||||
msgstr "Keine solche Metrik"
|
||||
|
||||
#, python-format
|
||||
msgid "No values returned by %(resource_id)s for %(metric_name)s."
|
||||
msgstr "Keine Werte von %(resource_id)s für %(metric_name)s zurückgegeben."
|
||||
|
||||
msgid "Noisy Neighbor"
|
||||
msgstr "Lauter Nachbar"
|
||||
|
||||
@@ -559,6 +555,9 @@ msgstr ""
|
||||
"Verhältnis der freigegebenen Compute-Knoten geteilt durch die Gesamtzahl der "
|
||||
"aktivierten Compute-Knoten."
|
||||
|
||||
msgid "Request not acceptable."
|
||||
msgstr "Anforderung nicht zulässig."
|
||||
|
||||
#, python-format
|
||||
msgid "Role name seems ambiguous: %s"
|
||||
msgstr "Der Rollenname scheint mehrdeutig: %s"
|
||||
@@ -648,10 +647,6 @@ msgstr "Die UUID oder der Name der Überprüfungsvorlage ist ungültig"
|
||||
msgid "The baremetal resource '%(name)s' could not be found"
|
||||
msgstr "Die Barmetal-Ressource '%(name)s' konnte nicht gefunden werden"
|
||||
|
||||
#, python-format
|
||||
msgid "The capacity %(capacity)s is not defined for '%(resource)s'"
|
||||
msgstr "Die Kapazität %(capacity)s ist nicht definiert für '%(resource)s'"
|
||||
|
||||
#, python-format
|
||||
msgid "The cluster data model '%(cdm)s' could not be built"
|
||||
msgstr "Das Clusterdatenmodell '%(cdm)s' konnte nicht erstellt werden"
|
||||
@@ -690,9 +685,6 @@ msgstr "Die Instanz '%(name)s' konnte nicht gefunden werden"
|
||||
msgid "The ironic node %(uuid)s could not be found"
|
||||
msgstr "Der Ironic Knoten %(uuid)s konnte nicht gefunden werden"
|
||||
|
||||
msgid "The metrics resource collector is not defined"
|
||||
msgstr "Der Metrikressourcen-Collector ist nicht definiert"
|
||||
|
||||
msgid "The number of VM migrations to be performed."
|
||||
msgstr "Die Anzahl der VM-Migrationen, die ausgeführt werden sollen."
|
||||
|
||||
|
||||
@@ -1,14 +1,15 @@
|
||||
# Andi Chandler <andi@gowling.com>, 2017. #zanata
|
||||
# Andi Chandler <andi@gowling.com>, 2018. #zanata
|
||||
# Andi Chandler <andi@gowling.com>, 2020. #zanata
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: watcher VERSION\n"
|
||||
"Report-Msgid-Bugs-To: https://bugs.launchpad.net/openstack-i18n/\n"
|
||||
"POT-Creation-Date: 2019-03-22 02:21+0000\n"
|
||||
"POT-Creation-Date: 2020-10-27 04:14+0000\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=UTF-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"PO-Revision-Date: 2018-11-07 06:14+0000\n"
|
||||
"PO-Revision-Date: 2020-10-28 11:02+0000\n"
|
||||
"Last-Translator: Andi Chandler <andi@gowling.com>\n"
|
||||
"Language-Team: English (United Kingdom)\n"
|
||||
"Language: en_GB\n"
|
||||
@@ -63,10 +64,6 @@ msgstr "'obj' argument type is not valid"
|
||||
msgid "'obj' argument type is not valid: %s"
|
||||
msgstr "'obj' argument type is not valid: %s"
|
||||
|
||||
#, python-format
|
||||
msgid "A datetime.datetime is required here. Got %s"
|
||||
msgstr "A datetime.datetime is required here. Got %s"
|
||||
|
||||
#, python-format
|
||||
msgid "A goal with UUID %(uuid)s already exists"
|
||||
msgstr "A goal with UUID %(uuid)s already exists"
|
||||
@@ -162,10 +159,6 @@ msgstr "An audit_template with UUID or name %(audit_template)s already exists"
|
||||
msgid "An indicator value should be a number"
|
||||
msgstr "An indicator value should be a number"
|
||||
|
||||
#, python-format
|
||||
msgid "An object of class %s is required here"
|
||||
msgstr "An object of class %s is required here"
|
||||
|
||||
msgid "An unknown exception occurred"
|
||||
msgstr "An unknown exception occurred"
|
||||
|
||||
@@ -195,19 +188,21 @@ msgstr "Audit Templates"
|
||||
msgid "Audit parameter %(parameter)s are not allowed"
|
||||
msgstr "Audit parameter %(parameter)s are not allowed"
|
||||
|
||||
#, python-format
|
||||
msgid "Audit state %(state)s is disallowed."
|
||||
msgstr "Audit state %(state)s is disallowed."
|
||||
|
||||
#, python-format
|
||||
msgid "Audit type %(audit_type)s could not be found"
|
||||
msgstr "Audit type %(audit_type)s could not be found"
|
||||
|
||||
#, python-format
|
||||
msgid "AuditTemplate %(audit_template)s could not be found"
|
||||
msgstr "AuditTemplate %(audit_template)s could not be found"
|
||||
msgid "Audit type %(audit_type)s is disallowed."
|
||||
msgstr "Audit type %(audit_type)s is disallowed."
|
||||
|
||||
#, python-format
|
||||
msgid ""
|
||||
"AuditTemplate %(audit_template)s is referenced by one or multiple audits"
|
||||
msgstr ""
|
||||
"AuditTemplate %(audit_template)s is referenced by one or multiple audits"
|
||||
msgid "AuditTemplate %(audit_template)s could not be found"
|
||||
msgstr "AuditTemplate %(audit_template)s could not be found"
|
||||
|
||||
msgid "Audits"
|
||||
msgstr "Audits"
|
||||
@@ -257,6 +252,9 @@ msgstr "Cannot overwrite UUID for an existing efficacy indicator."
|
||||
msgid "Cannot remove 'goal' attribute from an audit template"
|
||||
msgstr "Cannot remove 'goal' attribute from an audit template"
|
||||
|
||||
msgid "Ceilometer helper does not support statistic series method"
|
||||
msgstr "Ceilometer helper does not support statistic series method"
|
||||
|
||||
msgid "Cluster Maintaining"
|
||||
msgstr "Cluster Maintaining"
|
||||
|
||||
@@ -383,6 +381,9 @@ msgstr "Goal %(goal)s is invalid"
|
||||
msgid "Goals"
|
||||
msgstr "Goals"
|
||||
|
||||
msgid "Grafana helper does not support statistic series method"
|
||||
msgstr "Grafana helper does not support statistic series method"
|
||||
|
||||
msgid "Hardware Maintenance"
|
||||
msgstr "Hardware Maintenance"
|
||||
|
||||
@@ -439,9 +440,6 @@ msgstr "Invalid state: %(state)s"
|
||||
msgid "JSON list expected in feature argument"
|
||||
msgstr "JSON list expected in feature argument"
|
||||
|
||||
msgid "Keystone API endpoint is missing"
|
||||
msgstr "Keystone API endpoint is missing"
|
||||
|
||||
msgid "Limit must be positive"
|
||||
msgstr "Limit must be positive"
|
||||
|
||||
@@ -451,10 +449,17 @@ msgstr "Limit should be positive"
|
||||
msgid "Maximum time since last check-in for up service."
|
||||
msgstr "Maximum time since last check-in for up service."
|
||||
|
||||
#, python-format
|
||||
msgid "Metric: %(metric)s not available"
|
||||
msgstr "Metric: %(metric)s not available"
|
||||
|
||||
#, python-format
|
||||
msgid "Migration of type '%(migration_type)s' is not supported."
|
||||
msgstr "Migration of type '%(migration_type)s' is not supported."
|
||||
|
||||
msgid "Minimum Nova API Version"
|
||||
msgstr "Minimum Nova API Version"
|
||||
|
||||
msgid ""
|
||||
"Name of this node. This can be an opaque identifier. It is not necessarily a "
|
||||
"hostname, FQDN, or IP address. However, the node name must be valid within "
|
||||
@@ -468,19 +473,15 @@ msgstr ""
|
||||
msgid "No %(metric)s metric for %(host)s found."
|
||||
msgstr "No %(metric)s metric for %(host)s found."
|
||||
|
||||
msgid "No rows were returned"
|
||||
msgstr "No rows were returned"
|
||||
msgid "No datasources available"
|
||||
msgstr "No datasources available"
|
||||
|
||||
#, python-format
|
||||
msgid "No strategy could be found to achieve the '%(goal)s' goal."
|
||||
msgstr "No strategy could be found to achieve the '%(goal)s' goal."
|
||||
|
||||
msgid "No such metric"
|
||||
msgstr "No such metric"
|
||||
|
||||
#, python-format
|
||||
msgid "No values returned by %(resource_id)s for %(metric_name)s."
|
||||
msgstr "No values returned by %(resource_id)s for %(metric_name)s."
|
||||
msgid "Node Resource Consolidation strategy"
|
||||
msgstr "Node Resource Consolidation strategy"
|
||||
|
||||
msgid "Noisy Neighbor"
|
||||
msgstr "Noisy Neighbour"
|
||||
@@ -575,9 +576,6 @@ msgstr "Role name seems ambiguous: %s"
|
||||
msgid "Role not Found: %s"
|
||||
msgstr "Role not Found: %s"
|
||||
|
||||
msgid "Sample Check"
|
||||
msgstr "Sample Check"
|
||||
|
||||
msgid "Saving Energy"
|
||||
msgstr "Saving Energy"
|
||||
|
||||
@@ -636,6 +634,10 @@ msgstr "Strategy %(strategy)s could not be found"
|
||||
msgid "Strategy %(strategy)s is invalid"
|
||||
msgstr "Strategy %(strategy)s is invalid"
|
||||
|
||||
#, python-format
|
||||
msgid "The %(data_model_type)s data model could not be found"
|
||||
msgstr "The %(data_model_type)s data model could not be found"
|
||||
|
||||
#, python-format
|
||||
msgid "The %(name)s %(id)s could not be found"
|
||||
msgstr "The %(name)s %(id)s could not be found"
|
||||
@@ -667,10 +669,6 @@ msgstr "The audit template UUID or name specified is invalid"
|
||||
msgid "The baremetal resource '%(name)s' could not be found"
|
||||
msgstr "The baremetal resource '%(name)s' could not be found"
|
||||
|
||||
#, python-format
|
||||
msgid "The capacity %(capacity)s is not defined for '%(resource)s'"
|
||||
msgstr "The capacity %(capacity)s is not defined for '%(resource)s'"
|
||||
|
||||
#, python-format
|
||||
msgid "The cluster data model '%(cdm)s' could not be built"
|
||||
msgstr "The cluster data model '%(cdm)s' could not be built"
|
||||
@@ -709,8 +707,12 @@ msgstr "The instance '%(name)s' could not be found"
|
||||
msgid "The ironic node %(uuid)s could not be found"
|
||||
msgstr "The Ironic node %(uuid)s could not be found"
|
||||
|
||||
msgid "The metrics resource collector is not defined"
|
||||
msgstr "The metrics resource collector is not defined"
|
||||
#, python-format
|
||||
msgid "The mapped compute node for instance '%(uuid)s' could not be found."
|
||||
msgstr "The mapped compute node for instance '%(uuid)s' could not be found."
|
||||
|
||||
msgid "The node status is not defined"
|
||||
msgstr "The node status is not defined"
|
||||
|
||||
msgid "The number of VM migrations to be performed."
|
||||
msgstr "The number of VM migrations to be performed."
|
||||
@@ -775,6 +777,10 @@ msgstr "The total number of audited instances in strategy."
|
||||
msgid "The total number of enabled compute nodes."
|
||||
msgstr "The total number of enabled compute nodes."
|
||||
|
||||
#, python-format
|
||||
msgid "The value %(value)s for parameter %(parameter)s is invalid"
|
||||
msgstr "The value %(value)s for parameter %(parameter)s is invalid"
|
||||
|
||||
msgid "The value of original standard deviation."
|
||||
msgstr "The value of original standard deviation."
|
||||
|
||||
|
||||
@@ -15,7 +15,6 @@
|
||||
"""Utility methods for objects"""
|
||||
|
||||
import ast
|
||||
import six
|
||||
|
||||
from oslo_serialization import jsonutils
|
||||
from oslo_versionedobjects import fields
|
||||
@@ -67,7 +66,7 @@ class ListOfUUIDsField(fields.AutoTypedField):
|
||||
class FlexibleDict(fields.FieldType):
|
||||
@staticmethod
|
||||
def coerce(obj, attr, value):
|
||||
if isinstance(value, six.string_types):
|
||||
if isinstance(value, str):
|
||||
value = ast.literal_eval(value)
|
||||
return dict(value)
|
||||
|
||||
@@ -87,7 +86,7 @@ class FlexibleDictField(fields.AutoTypedField):
|
||||
class FlexibleListOfDict(fields.FieldType):
|
||||
@staticmethod
|
||||
def coerce(obj, attr, value):
|
||||
if isinstance(value, six.string_types):
|
||||
if isinstance(value, str):
|
||||
value = ast.literal_eval(value)
|
||||
return list(value)
|
||||
|
||||
@@ -106,7 +105,7 @@ class FlexibleListOfDictField(fields.AutoTypedField):
|
||||
|
||||
class Json(fields.FieldType):
|
||||
def coerce(self, obj, attr, value):
|
||||
if isinstance(value, six.string_types):
|
||||
if isinstance(value, str):
|
||||
loaded = jsonutils.loads(value)
|
||||
return loaded
|
||||
return value
|
||||
|
||||
@@ -23,12 +23,12 @@
|
||||
# NOTE(deva): import auth_token so we can override a config option
|
||||
|
||||
import copy
|
||||
import mock
|
||||
from unittest import mock
|
||||
from urllib import parse as urlparse
|
||||
|
||||
from oslo_config import cfg
|
||||
import pecan
|
||||
import pecan.testing
|
||||
from six.moves.urllib import parse as urlparse
|
||||
|
||||
from watcher.api import hooks
|
||||
from watcher.common import context as watcher_context
|
||||
|
||||
@@ -12,7 +12,7 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import imp
|
||||
import importlib
|
||||
from oslo_config import cfg
|
||||
from watcher.api import config as api_config
|
||||
from watcher.tests.api import base
|
||||
@@ -23,13 +23,13 @@ class TestRoot(base.FunctionalTest):
|
||||
def test_config_enable_webhooks_auth(self):
|
||||
acl_public_routes = ['/']
|
||||
cfg.CONF.set_override('enable_webhooks_auth', True, 'api')
|
||||
imp.reload(api_config)
|
||||
importlib.reload(api_config)
|
||||
self.assertEqual(acl_public_routes,
|
||||
api_config.app['acl_public_routes'])
|
||||
|
||||
def test_config_disable_webhooks_auth(self):
|
||||
acl_public_routes = ['/', '/v1/webhooks/.*']
|
||||
cfg.CONF.set_override('enable_webhooks_auth', False, 'api')
|
||||
imp.reload(api_config)
|
||||
importlib.reload(api_config)
|
||||
self.assertEqual(acl_public_routes,
|
||||
api_config.app['acl_public_routes'])
|
||||
|
||||
@@ -14,15 +14,11 @@
|
||||
|
||||
"""Tests for the Pecan API hooks."""
|
||||
|
||||
from __future__ import unicode_literals
|
||||
|
||||
import mock
|
||||
from http import client as http_client
|
||||
from oslo_config import cfg
|
||||
import oslo_messaging as messaging
|
||||
from oslo_serialization import jsonutils
|
||||
import six
|
||||
from six.moves import http_client
|
||||
|
||||
from unittest import mock
|
||||
from watcher.api.controllers import root
|
||||
from watcher.api import hooks
|
||||
from watcher.common import context
|
||||
@@ -145,7 +141,7 @@ class TestNoExceptionTracebackHook(base.FunctionalTest):
|
||||
# we don't care about this garbage.
|
||||
expected_msg = ("Remote error: %s %s"
|
||||
% (test_exc_type, self.MSG_WITHOUT_TRACE) +
|
||||
("\n[u'" if six.PY2 else "\n['"))
|
||||
"\n['")
|
||||
actual_msg = jsonutils.loads(
|
||||
response.json['error_message'])['faultstring']
|
||||
self.assertEqual(expected_msg, actual_msg)
|
||||
|
||||
@@ -15,7 +15,7 @@
|
||||
from apscheduler.schedulers import background
|
||||
import datetime
|
||||
import freezegun
|
||||
import mock
|
||||
from unittest import mock
|
||||
|
||||
from watcher.api import scheduling
|
||||
from watcher.notifications import service
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user