# -*- coding: utf-8 -*-
# vim: tabstop=4 shiftwidth=4 softtabstop=4
#
# Copyright (C) 2010-2023 GEM Foundation
#
# OpenQuake is free software: you can redistribute it and/or modify it
# under the terms of the GNU Affero General Public License as published
# by the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# OpenQuake is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with OpenQuake. If not, see <http://www.gnu.org/licenses/>.
"""Engine: A collection of fundamental functions for initializing and running
calculations."""
import os
import re
import sys
import copy
import json
import time
import pickle
import socket
import signal
import getpass
import logging
import itertools
import platform
from os.path import getsize
from datetime import datetime
import psutil
import h5py
import numpy
try:
from setproctitle import setproctitle
except ImportError:
def setproctitle(title):
"Do nothing"
from urllib.request import urlopen, Request
from openquake.baselib.python3compat import decode
from openquake.baselib import parallel, general, config, workerpool as w
from openquake.hazardlib import valid
from openquake.commonlib.oqvalidation import OqParam
from openquake.commonlib import readinput
from openquake.calculators import base, export
from openquake.commonlib import logs
USER = getpass.getuser()
OQ_API = 'https://api.openquake.org'
OQ_DISTRIBUTE = parallel.oq_distribute()
MB = 1024 ** 2
_PID = os.getpid() # the PID
_PPID = os.getppid() # the controlling terminal PID
GET_JOBS = '''--- executing or submitted
SELECT * FROM job WHERE status IN ('executing', 'submitted')
AND host=?x AND is_running=1 AND pid > 0 ORDER BY id'''
[docs]def workers_stop():
w.WorkerMaster(config.zworkers).stop()
[docs]def get_zmq_ports():
"""
:returns: an array with the receiver ports
"""
start, stop = config.dbserver.receiver_ports.split('-')
return numpy.arange(int(start), int(stop))
[docs]def set_concurrent_tasks_default(calc):
"""
Look at the number of available workers and update the parameter
OqParam.concurrent_tasks.default. Abort the calculations if no
workers are available. Do nothing for trivial distributions.
"""
if OQ_DISTRIBUTE in 'no processpool ipp': # do nothing
num_workers = 0 if OQ_DISTRIBUTE == 'no' else parallel.Starmap.CT // 2
logging.warning('Using %d cores on %s', num_workers, platform.node())
return
master = w.WorkerMaster(config.zworkers)
num_workers = sum(total for host, running, total in master.wait())
if num_workers == 0:
logging.critical("No live compute nodes, aborting calculation")
logs.dbcmd('finish', calc.datastore.calc_id, 'failed')
sys.exit(1)
parallel.Starmap.CT = num_workers * 2
OqParam.concurrent_tasks.default = num_workers * 2
logging.warning('Using %d %s workers', num_workers, OQ_DISTRIBUTE)
[docs]def expose_outputs(dstore, owner=USER, status='complete'):
"""
Build a correspondence between the outputs in the datastore and the
ones in the database.
:param dstore: datastore
"""
oq = dstore['oqparam']
exportable = set(ekey[0] for ekey in export.export)
calcmode = oq.calculation_mode
dskeys = set(dstore) & exportable # exportable datastore keys
dskeys.add('fullreport')
if 'avg_gmf' in dskeys:
dskeys.remove('avg_gmf') # hide
rlzs = dstore['full_lt'].rlzs
if len(rlzs) > 1:
dskeys.add('realizations')
hdf5 = dstore.hdf5
if 'hcurves-stats' in hdf5 or 'hcurves-rlzs' in hdf5:
if oq.hazard_stats() or oq.individual_rlzs or len(rlzs) == 1:
dskeys.add('hcurves')
if oq.uniform_hazard_spectra:
dskeys.add('uhs') # export them
if oq.hazard_maps:
dskeys.add('hmaps') # export them
if len(rlzs) > 1 and not oq.collect_rlzs:
if 'aggrisk' in dstore:
dskeys.add('aggrisk-stats')
if 'aggcurves' in dstore:
dskeys.add('aggcurves-stats')
if not oq.individual_rlzs:
for out in ['avg_losses-rlzs', 'aggrisk', 'aggcurves']:
if out in dskeys:
dskeys.remove(out)
if 'curves-rlzs' in dstore and len(rlzs) == 1:
dskeys.add('loss_curves-rlzs')
if 'curves-stats' in dstore and len(rlzs) > 1:
dskeys.add('loss_curves-stats')
if oq.conditional_loss_poes: # expose loss_maps outputs
if 'loss_curves-stats' in dstore:
dskeys.add('loss_maps-stats')
if 'ruptures' in dskeys:
if 'scenario' in calcmode:
# do not export, as requested by Vitor
exportable.remove('ruptures')
else:
dskeys.add('event_based_mfd')
if 'hmaps' in dskeys and not oq.hazard_maps:
dskeys.remove('hmaps') # do not export the hazard maps
if logs.dbcmd('get_job', dstore.calc_id) is None:
# the calculation has not been imported in the db yet
logs.dbcmd('import_job', dstore.calc_id, oq.calculation_mode,
oq.description + ' [parent]', owner, status,
oq.hazard_calculation_id, dstore.datadir)
keysize = []
for key in sorted(dskeys & exportable):
try:
size_mb = dstore.getsize(key) / MB
except (KeyError, AttributeError):
size_mb = -1
if size_mb:
keysize.append((key, size_mb))
ds_size = dstore.getsize() / MB
logs.dbcmd('create_outputs', dstore.calc_id, keysize, ds_size)
[docs]class MasterKilled(KeyboardInterrupt):
"Exception raised when a job is killed manually"
[docs]def inhibitSigInt(signum, _stack):
logging.warning('Killing job, please wait')
[docs]def manage_signals(signum, _stack):
"""
Convert a SIGTERM into a SystemExit exception and a SIGINT/SIGHUP into
a MasterKilled exception with an appropriate error message.
:param int signum: the number of the received signal
:param _stack: the current frame object, ignored
"""
if signum == signal.SIGINT:
workers_stop()
raise MasterKilled('The openquake master process was killed manually')
if signum == signal.SIGTERM:
workers_stop()
raise SystemExit('Terminated')
if hasattr(signal, 'SIGHUP'): # there is no SIGHUP on Windows
# kill the calculation only if os.getppid() != _PPID, i.e. the
# controlling terminal died; in the workers, do nothing
if signum == signal.SIGHUP and os.getppid() != _PPID:
workers_stop()
raise MasterKilled(
'The openquake master lost its controlling terminal')
[docs]def register_signals():
# register the manage_signals callback for SIGTERM, SIGINT, SIGHUP
# when using the Django development server this module is imported by a
# thread, so one gets a `ValueError: signal only works in main thread` that
# can be safely ignored
try:
signal.signal(signal.SIGTERM, manage_signals)
signal.signal(signal.SIGINT, manage_signals)
if hasattr(signal, 'SIGHUP'):
# Do not register our SIGHUP handler if running with 'nohup'
if signal.getsignal(signal.SIGHUP) != signal.SIG_IGN:
signal.signal(signal.SIGHUP, manage_signals)
except ValueError:
pass
[docs]def poll_queue(job_id, poll_time):
"""
Check the queue of executing/submitted jobs and exit when there is
a free slot.
"""
try:
host = socket.gethostname()
except Exception: # gaierror
host = None
offset = config.distribution.serialize_jobs - 1
if offset >= 0:
first_time = True
while True:
running = logs.dbcmd(GET_JOBS, host)
previous = [job.id for job in running if job.id < job_id - offset]
if previous:
if first_time:
logs.dbcmd('update_job', job_id,
{'status': 'submitted', 'pid': _PID})
first_time = False
# the logging is not yet initialized, so use a print
print('Waiting for jobs %s' % ' '.join(map(str, previous)))
time.sleep(poll_time)
else:
break
[docs]def run_calc(log):
"""
Run a calculation.
:param log:
LogContext of the current job
"""
register_signals()
setproctitle('oq-job-%d' % log.calc_id)
with log:
# check the available memory before starting
while True:
used_mem = psutil.virtual_memory().percent
if used_mem < 80: # continue if little memory is in use
break
logging.info('Memory occupation %d%%, the user should free '
'some memory', used_mem)
time.sleep(5)
oqparam = log.get_oqparam()
calc = base.calculators(oqparam, log.calc_id)
try:
hostname = socket.gethostname()
except Exception: # gaierror
hostname = 'localhost'
logging.info('%s@%s running %s [--hc=%s]',
USER,
hostname,
calc.oqparam.inputs['job_ini'],
calc.oqparam.hazard_calculation_id)
msg = check_obsolete_version(oqparam.calculation_mode)
# NB: disabling the warning should be done only for users with
# an updated LTS version, but we are doing it for all users
# if msg:
# logging.warning(msg)
calc.from_engine = True
if config.zworkers['host_cores']:
set_concurrent_tasks_default(calc)
else:
logging.warning('Assuming %d %s workers',
parallel.Starmap.CT // 2, OQ_DISTRIBUTE)
t0 = time.time()
calc.run(shutdown=True)
logging.info('Exposing the outputs to the database')
expose_outputs(calc.datastore)
path = calc.datastore.filename
size = general.humansize(getsize(path))
logging.info('Stored %s on %s in %d seconds',
size, path, time.time() - t0)
calc.datastore.close()
for line in logs.dbcmd('list_outputs', log.calc_id, False):
general.safeprint(line)
# sanity check to make sure that the logging on file is working
if (log.log_file and log.log_file != os.devnull and
getsize(log.log_file) == 0):
logging.warning('The log file %s is empty!?' % log.log_file)
return calc
[docs]def create_jobs(job_inis, log_level=logging.INFO, log_file=None,
user_name=USER, hc_id=None, multi=True, host=None):
"""
Create job records on the database.
:param job_inis: a list of pathnames or a list of dictionaries
:returns: a list of LogContext objects
"""
try:
host = socket.gethostname()
except Exception: # gaierror
host = None
if len(job_inis) > 1 and not hc_id and not multi: # first job as hc
job = logs.init("job", job_inis[0], log_level, log_file,
user_name, hc_id, host)
hc_id = job.calc_id
jobs = [job]
job_inis = job_inis[1:]
else:
jobs = []
for job_ini in job_inis:
if isinstance(job_ini, dict):
dic = job_ini
else:
# NB: `get_params` must NOT log, since the logging is not
# configured yet, otherwise the log will disappear :-(
dic = readinput.get_params(job_ini)
dic['hazard_calculation_id'] = hc_id
if 'sensitivity_analysis' in dic:
# this part is tested in commands_test and in oq-risk-tests
analysis = valid.dictionary(dic['sensitivity_analysis'])
for values in itertools.product(*analysis.values()):
jobdic = copy.deepcopy(dic)
pars = dict(zip(analysis, map(str, values)))
readinput.update(jobdic, pars.items(), dic['base_path'])
jobdic['description'] = '%s %s' % (dic['description'], pars)
new = logs.init('job', jobdic, log_level, log_file,
user_name, hc_id, host)
jobs.append(new)
else:
jobs.append(logs.init('job', dic, log_level, log_file,
user_name, hc_id, host))
if multi:
for job in jobs:
job.multi = True
return jobs
[docs]def cleanup(kind):
"""
Stop or kill the zmq workers if serialize_jobs == 1.
"""
assert kind in ("stop", "kill"), kind
if OQ_DISTRIBUTE != 'zmq' or config.distribution.serialize_jobs > 1:
return # do nothing
if kind == 'stop':
# called in the regular case, does not require ssh access
print('Stopping the workers')
workers_stop()
elif kind == 'kill':
# called in case of exceptions (or out of memory), requires ssh
print('Asking the DbServer to kill the workers')
logs.dbcmd('workers_kill', config.zworkers)
[docs]def run_jobs(jobctxs, concurrent_jobs=3):
"""
Run jobs using the specified config file and other options.
:param jobctxs:
List of LogContexts
:param concurrent_jobs:
How many jobs to run concurrently (default 3)
"""
hc_id = jobctxs[-1].params['hazard_calculation_id']
if hc_id:
job = logs.dbcmd('get_job', hc_id)
ppath = job.ds_calc_dir + '.hdf5'
if os.path.exists(ppath):
version = logs.dbcmd('engine_version')
with h5py.File(ppath, 'r') as f:
prev_version = f.attrs['engine_version']
if prev_version != version:
# here the logger is not initialized yet
print('Starting from a hazard (%d) computed with'
' an obsolete version of the engine: %s' %
(hc_id, version))
jobarray = len(jobctxs) > 1 and jobctxs[0].multi
try:
poll_queue(jobctxs[0].calc_id, poll_time=15)
# wait for an empty slot or a CTRL-C
except BaseException:
# the job aborted even before starting
for job in jobctxs:
logs.dbcmd('finish', job.calc_id, 'aborted')
return jobctxs
for job in jobctxs:
dic = {'status': 'executing', 'pid': _PID,
'start_time': datetime.utcnow()}
logs.dbcmd('update_job', job.calc_id, dic)
try:
if OQ_DISTRIBUTE == 'zmq' and w.WorkerMaster(
config.zworkers).status() == []:
print('Asking the DbServer to start the workers %s' %
config.zworkers.host_cores)
logs.dbcmd('workers_start', config.zworkers) # start the workers
allargs = [(ctx,) for ctx in jobctxs]
if jobarray and OQ_DISTRIBUTE != 'no':
parallel.multispawn(run_calc, allargs, concurrent_jobs)
else:
for jobctx in jobctxs:
run_calc(jobctx)
cleanup('stop')
except Exception:
ids = [jc.calc_id for jc in jobctxs]
rows = logs.dbcmd("SELECT id FROM job WHERE id IN (?X) "
"AND status IN ('created', 'executing')", ids)
for job_id, in rows:
logs.dbcmd("set_status", job_id, 'failed')
cleanup('kill')
raise
return jobctxs
[docs]def version_triple(tag):
"""
returns: a triple of integers from a version tag
"""
groups = re.match(r'v?(\d+)\.(\d+)\.(\d+)', tag).groups()
return tuple(int(n) for n in groups)
[docs]def check_obsolete_version(calculation_mode='WebUI'):
"""
Check if there is a newer version of the engine.
:param calculation_mode:
- the calculation mode when called from the engine
- an empty string when called from the WebUI
:returns:
- a message if the running version of the engine is obsolete
- the empty string if the engine is updated
- None if the check could not be performed (i.e. github is down)
"""
if os.environ.get('JENKINS_URL') or os.environ.get('CI'):
# avoid flooding our API server with requests from CI systems
return
version = logs.dbcmd('engine_version')
logging.info('Using engine version %s', version)
headers = {'User-Agent': 'OpenQuake Engine %s;%s;%s;%s' %
(version, calculation_mode, platform.platform(),
config.distribution.oq_distribute)}
try:
req = Request(OQ_API + '/engine/latest', headers=headers)
# NB: a timeout < 1 does not work
data = urlopen(req, timeout=1).read() # bytes
tag_name = json.loads(decode(data))['tag_name']
current = version_triple(version)
latest = version_triple(tag_name)
except Exception: # page not available or wrong version tag
msg = ('An error occurred while calling %s/engine/latest to check'
' if the installed version of the engine is up to date.' %
OQ_API)
logging.warning(msg)
return
if current < latest:
return ('Version %s of the engine is available, but you are '
'still using version %s' % (tag_name, version))
else:
return ''
if __name__ == '__main__':
from openquake.server import dbserver
# run a LogContext object stored in a pickle file, called by job.yaml
with open(sys.argv[1], 'rb') as f:
jobctx = pickle.load(f)
dbserver.ensure_on()
run_jobs([jobctx])