Compare commits
9 Commits
stable/202
...
stable/202
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
8f8d537330 | ||
|
|
6264f17c92 | ||
|
|
4fca7d246a | ||
|
|
64ba589f80 | ||
|
|
accc7a2a22 | ||
|
|
70d92a75cc | ||
|
|
f5ef5557ab | ||
|
|
c604e00b4d | ||
|
|
6b433b3547 |
@@ -2,4 +2,4 @@
|
||||
host=review.opendev.org
|
||||
port=29418
|
||||
project=openstack/watcher.git
|
||||
defaultbranch=stable/2024.2
|
||||
defaultbranch=stable/2024.1
|
||||
|
||||
@@ -275,9 +275,6 @@ function install_watcherclient {
|
||||
git_clone_by_name "python-watcherclient"
|
||||
setup_dev_lib "python-watcherclient"
|
||||
fi
|
||||
if [[ "$GLOBAL_VENV" == "True" ]]; then
|
||||
sudo ln -sf /opt/stack/data/venv/bin/watcher /usr/local/bin
|
||||
fi
|
||||
}
|
||||
|
||||
# install_watcher() - Collect source and prepare
|
||||
|
||||
@@ -70,7 +70,7 @@ then write_uwsgi_config "$WATCHER_UWSGI_CONF" "$WATCHER_UWSGI" "/infra-optim"
|
||||
fi
|
||||
|
||||
# Migrate the database
|
||||
$WATCHER_BIN_DIR/watcher-db-manage upgrade || die $LINO "DB migration error"
|
||||
watcher-db-manage upgrade || die $LINO "DB migration error"
|
||||
|
||||
start_watcher
|
||||
|
||||
|
||||
@@ -3,16 +3,15 @@
|
||||
# Andi Chandler <andi@gowling.com>, 2020. #zanata
|
||||
# Andi Chandler <andi@gowling.com>, 2022. #zanata
|
||||
# Andi Chandler <andi@gowling.com>, 2023. #zanata
|
||||
# Andi Chandler <andi@gowling.com>, 2024. #zanata
|
||||
msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: python-watcher\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2024-05-31 14:40+0000\n"
|
||||
"POT-Creation-Date: 2023-08-14 03:05+0000\n"
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=UTF-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"PO-Revision-Date: 2024-04-18 12:21+0000\n"
|
||||
"PO-Revision-Date: 2023-06-21 07:54+0000\n"
|
||||
"Last-Translator: Andi Chandler <andi@gowling.com>\n"
|
||||
"Language-Team: English (United Kingdom)\n"
|
||||
"Language: en_GB\n"
|
||||
@@ -64,9 +63,6 @@ msgstr "2.0.0"
|
||||
msgid "2023.1 Series Release Notes"
|
||||
msgstr "2023.1 Series Release Notes"
|
||||
|
||||
msgid "2023.2 Series Release Notes"
|
||||
msgstr "2023.2 Series Release Notes"
|
||||
|
||||
msgid "3.0.0"
|
||||
msgstr "3.0.0"
|
||||
|
||||
|
||||
2
tox.ini
2
tox.ini
@@ -8,7 +8,7 @@ basepython = python3
|
||||
usedevelop = True
|
||||
allowlist_externals = find
|
||||
rm
|
||||
install_command = pip install -c{env:TOX_CONSTRAINTS_FILE:https://releases.openstack.org/constraints/upper/2024.2} {opts} {packages}
|
||||
install_command = pip install -c{env:TOX_CONSTRAINTS_FILE:https://releases.openstack.org/constraints/upper/2024.1} {opts} {packages}
|
||||
setenv =
|
||||
VIRTUAL_ENV={envdir}
|
||||
deps =
|
||||
|
||||
@@ -28,7 +28,7 @@ def upgrade():
|
||||
|
||||
op.create_table(
|
||||
'apscheduler_jobs',
|
||||
sa.Column('id', sa.Unicode(191),
|
||||
sa.Column('id', sa.Unicode(191, _warn_on_bytestring=False),
|
||||
nullable=False),
|
||||
sa.Column('next_run_time', sa.Float(25), index=True),
|
||||
sa.Column('job_state', sa.LargeBinary, nullable=False),
|
||||
|
||||
@@ -228,8 +228,7 @@ class Connection(api.BaseConnection):
|
||||
for relationship in relationships:
|
||||
if not relationship.uselist:
|
||||
# We have a One-to-X relationship
|
||||
query = query.options(joinedload(
|
||||
getattr(model, relationship.key)))
|
||||
query = query.options(joinedload(relationship.key))
|
||||
return query
|
||||
|
||||
@oslo_db_api.retry_on_deadlock
|
||||
|
||||
@@ -22,7 +22,6 @@ from apscheduler.jobstores.base import ConflictingIdError
|
||||
from apscheduler.jobstores import sqlalchemy
|
||||
from apscheduler.util import datetime_to_utc_timestamp
|
||||
from apscheduler.util import maybe_ref
|
||||
from apscheduler.util import utc_timestamp_to_datetime
|
||||
|
||||
from watcher.common import context
|
||||
from watcher.common import service
|
||||
@@ -33,7 +32,7 @@ try:
|
||||
except ImportError: # pragma: nocover
|
||||
import pickle
|
||||
|
||||
from sqlalchemy import Table, MetaData, select, and_, null
|
||||
from sqlalchemy import Table, MetaData, select, and_
|
||||
from sqlalchemy.exc import IntegrityError
|
||||
|
||||
|
||||
@@ -59,7 +58,8 @@ class WatcherJobStore(sqlalchemy.SQLAlchemyJobStore):
|
||||
super(WatcherJobStore, self).__init__(url, engine, tablename,
|
||||
metadata, pickle_protocol)
|
||||
metadata = maybe_ref(metadata) or MetaData()
|
||||
self.jobs_t = Table(tablename, metadata, autoload_with=engine)
|
||||
self.jobs_t = Table(tablename, metadata, autoload=True,
|
||||
autoload_with=engine)
|
||||
service_ident = service.ServiceHeartbeat.get_service_name()
|
||||
self.tag = tag or {'host': service_ident[0], 'name': service_ident[1]}
|
||||
self.service_id = objects.Service.list(context=context.make_context(),
|
||||
@@ -79,8 +79,7 @@ class WatcherJobStore(sqlalchemy.SQLAlchemyJobStore):
|
||||
'tag': jsonutils.dumps(self.tag)
|
||||
})
|
||||
try:
|
||||
with self.engine.begin() as conn:
|
||||
conn.execute(insert)
|
||||
self.engine.execute(insert)
|
||||
except IntegrityError:
|
||||
raise ConflictingIdError(job.id)
|
||||
|
||||
@@ -89,36 +88,20 @@ class WatcherJobStore(sqlalchemy.SQLAlchemyJobStore):
|
||||
self._fix_paused_jobs_sorting(jobs)
|
||||
return jobs
|
||||
|
||||
def get_next_run_time(self):
|
||||
selectable = select(self.jobs_t.c.next_run_time).\
|
||||
where(self.jobs_t.c.next_run_time != null()).\
|
||||
order_by(self.jobs_t.c.next_run_time).limit(1)
|
||||
with self.engine.begin() as connection:
|
||||
# NOTE(danms): The apscheduler implementation of this gets a
|
||||
# decimal.Decimal back from scalar() which causes
|
||||
# utc_timestamp_to_datetime() to choke since it is expecting a
|
||||
# python float. Assume this is SQLAlchemy 2.0 stuff, so just
|
||||
# coerce to a float here.
|
||||
next_run_time = connection.execute(selectable).scalar()
|
||||
return utc_timestamp_to_datetime(float(next_run_time)
|
||||
if next_run_time is not None
|
||||
else None)
|
||||
|
||||
def _get_jobs(self, *conditions):
|
||||
jobs = []
|
||||
conditions += (self.jobs_t.c.service_id == self.service_id,)
|
||||
selectable = select(
|
||||
self.jobs_t.c.id, self.jobs_t.c.job_state, self.jobs_t.c.tag
|
||||
[self.jobs_t.c.id, self.jobs_t.c.job_state, self.jobs_t.c.tag]
|
||||
).order_by(self.jobs_t.c.next_run_time).where(and_(*conditions))
|
||||
failed_job_ids = set()
|
||||
with self.engine.begin() as conn:
|
||||
for row in conn.execute(selectable):
|
||||
try:
|
||||
jobs.append(self._reconstitute_job(row.job_state))
|
||||
except Exception:
|
||||
self._logger.exception(
|
||||
'Unable to restore job "%s" -- removing it', row.id)
|
||||
failed_job_ids.add(row.id)
|
||||
for row in self.engine.execute(selectable):
|
||||
try:
|
||||
jobs.append(self._reconstitute_job(row.job_state))
|
||||
except Exception:
|
||||
self._logger.exception(
|
||||
'Unable to restore job "%s" -- removing it', row.id)
|
||||
failed_job_ids.add(row.id)
|
||||
|
||||
# Remove all the jobs we failed to restore
|
||||
if failed_job_ids:
|
||||
|
||||
Reference in New Issue
Block a user