# -*- coding: utf-8 -*-
# vim: tabstop=4 shiftwidth=4 softtabstop=4
#
# Copyright (C) 2015-2023 GEM Foundation
#
# OpenQuake is free software: you can redistribute it and/or modify it
# under the terms of the GNU Affero General Public License as published
# by the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# OpenQuake is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with OpenQuake. If not, see <http://www.gnu.org/licenses/>.
import os
import operator
import collections
import pickle
import toml
import copy
import logging
from dataclasses import dataclass
import numpy
from openquake.hazardlib.source.multi_fault import MultiFaultSource
from openquake.baselib import hdf5
from openquake.baselib.general import groupby, block_splitter
from openquake.baselib.node import context, striptag, Node, node_to_dict
from openquake.hazardlib import geo, mfd, pmf, source, tom, valid, InvalidFile
from openquake.hazardlib.tom import PoissonTOM
from openquake.hazardlib.source import NonParametricSeismicSource
U32 = numpy.uint32
F32 = numpy.float32
F64 = numpy.float64
EPSILON = 1E-12
source_dt = numpy.dtype([('source_id', U32), ('num_ruptures', U32),
('pik', hdf5.vuint8)])
KNOWN_MFDS = ('incrementalMFD', 'truncGutenbergRichterMFD',
'arbitraryMFD', 'YoungsCoppersmithMFD', 'multiMFD',
'taperedGutenbergRichterMFD')
EXCLUDE_FROM_GEOM_PROPS = (
'Polygon', 'Point', 'MultiPoint', 'LineString', '3D MultiLineString',
'3D MultiPolygon', 'posList')
[docs]def fix_dupl(dist, fname=None, lineno=None):
"""
Fix the distribution if it contains identical values or raise an error.
:param dist:
a list of pairs [(prob, value)...] for a hypocenter or nodal plane dist
:param fname:
the file which is being read; if it is None, it means you are writing
the distribution: in that case raise an error for duplicated values
:param lineno:
the line number of the file which is being read (None in writing mode)
"""
n = len(dist)
values = collections.defaultdict(float) # dict value -> probability
# value can be a scalar (hypocenter depth) or a triple
# (strike, dip, rake) for a nodal plane distribution
got = []
for prob, value in dist:
if prob == 0:
raise ValueError('Zero probability in subnode %s' % value)
values[value] += prob
got.append(value)
if len(values) < n:
if fname is None: # when called from the sourcewriter
raise ValueError('There are repeated values in %s' % got)
else:
logging.warning('There were repeated values %s in %s:%s',
extract_dupl(got), fname, lineno)
assert abs(sum(values.values()) - 1) < EPSILON # sanity check
newdist = sorted([(p, v) for v, p in values.items()])
if isinstance(newdist[0][1], tuple): # nodal planes
newdist = [(p, geo.nodalplane.NodalPlane(*v))
for p, v in newdist]
# run hazardlib/tests/data/context/job.ini to check this;
# you will get [(0.2, 6.0), (0.2, 8.0), (0.2, 10.0), (0.4, 2.0)]
dist[:] = newdist
[docs]def rounded_unique(mags, idxs):
"""
:param mags: a list of magnitudes
:param idxs: a list of tuples of section indices
:returns: an array of magnitudes rounded to 2 digits
:raises: ValueError if the rounded magnitudes contain duplicates
"""
mags = numpy.round(mags, 2)
mag_idxs = [(mag, ' '.join(idx)) for mag, idx in zip(mags, idxs)]
dupl = extract_dupl(mag_idxs)
if dupl:
logging.error('the pair (mag=%s, idxs=%s) is duplicated' % dupl[0])
return mags
[docs]class SourceGroup(collections.abc.Sequence):
"""
A container for the following parameters:
:param str trt:
the tectonic region type all the sources belong to
:param list sources:
a list of hazardlib source objects
:param name:
The name of the group
:param src_interdep:
A string specifying if the sources in this cluster are independent or
mutually exclusive
:param rup_indep:
A string specifying if the ruptures within each source of the cluster
are independent or mutually exclusive
:param weights:
A dictionary whose keys are the source IDs of the cluster and the
values are the weights associated with each source
:param min_mag:
the minimum magnitude among the given sources
:param max_mag:
the maximum magnitude among the given sources
:param id:
an optional numeric ID (default 0) set by the engine and used
when serializing SourceModels to HDF5
:param temporal_occurrence_model:
A temporal occurrence model controlling the source group occurrence
:param cluster:
A boolean indicating if the sources behaves as a cluster similarly
to what used by the USGS for the New Madrid in the 2008 National
Hazard Model.
"""
changes = 0 # set in apply_uncertainty
[docs] @classmethod
def collect(cls, sources):
"""
:param sources: dictionaries with a key 'tectonicRegion'
:returns: an ordered list of SourceGroup instances
"""
source_stats_dict = {}
for src in sources:
trt = src['tectonicRegion']
if trt not in source_stats_dict:
source_stats_dict[trt] = SourceGroup(trt)
sg = source_stats_dict[trt]
if not sg.sources:
# we append just one source per SourceGroup, so that
# the memory occupation is insignificant
sg.sources.append(src)
# return SourceGroups, ordered by TRT string
return sorted(source_stats_dict.values())
def __init__(self, trt, sources=None, name=None, src_interdep='indep',
rup_interdep='indep', grp_probability=1.,
min_mag={'default': 0}, max_mag=None,
temporal_occurrence_model=None, cluster=False):
# checks
self.trt = trt
self.sources = []
self.name = name
self.src_interdep = src_interdep
self.rup_interdep = rup_interdep
self._check_init_variables(sources, name, src_interdep, rup_interdep)
self.grp_probability = grp_probability
self.min_mag = min_mag
self.max_mag = max_mag
if sources:
for src in sorted(sources, key=operator.attrgetter('source_id')):
self.update(src)
self.source_model = None # to be set later, in FullLogicTree
self.temporal_occurrence_model = temporal_occurrence_model
self.cluster = cluster
# check weights in case of mutually exclusive ruptures
if rup_interdep == 'mutex':
for src in self.sources:
assert isinstance(src, NonParametricSeismicSource)
for rup, _ in src.data:
assert rup.weight is not None
@property
def tom_name(self):
"""
:returns: name of the associated temporal occurrence model
"""
if self.temporal_occurrence_model:
return self.temporal_occurrence_model.__class__.__name__
else:
return 'PoissonTOM'
@property
def atomic(self):
"""
:returns: True if the group cannot be split
"""
return (self.cluster or self.src_interdep == 'mutex' or
self.rup_interdep == 'mutex')
@property
def weight(self):
"""
:returns: total weight of the underlying sources
"""
return sum(src.weight for src in self)
def _check_init_variables(self, src_list, name,
src_interdep, rup_interdep):
if src_interdep not in ('indep', 'mutex'):
raise ValueError('source interdependence incorrect %s ' %
src_interdep)
if rup_interdep not in ('indep', 'mutex'):
raise ValueError('rupture interdependence incorrect %s ' %
rup_interdep)
# check TRT
if src_list: # can be None
for src in src_list:
assert src.tectonic_region_type == self.trt, (
src.tectonic_region_type, self.trt)
# Mutually exclusive ruptures can only belong to non-parametric
# sources
if rup_interdep == 'mutex':
if not isinstance(src, NonParametricSeismicSource):
msg = "Mutually exclusive ruptures can only be "
msg += "modelled using non-parametric sources"
raise ValueError(msg)
[docs] def update(self, src):
"""
Update the attributes sources, min_mag, max_mag
according to the given source.
:param src:
an instance of :class:
`openquake.hazardlib.source.base.BaseSeismicSource`
"""
assert src.tectonic_region_type == self.trt, (
src.tectonic_region_type, self.trt)
src.min_mag = max(src.min_mag, self.min_mag.get(self.trt)
or self.min_mag['default'])
if hasattr(src, 'mags'):
# honor minimum_magnitude for multifault sources
ok = src.mags >= src.min_mag
for attr in 'mags rupture_idxs probs_occur rakes'.split():
setattr(src, attr, numpy.array(getattr(src, attr))[ok])
if src.min_mag and not src.get_mags(): # filtered out
return
# checking mutex ruptures
if (not isinstance(src, NonParametricSeismicSource) and
self.rup_interdep == 'mutex'):
msg = "Mutually exclusive ruptures can only be "
msg += "modelled using non-parametric sources"
raise ValueError(msg)
self.sources.append(src)
_, max_mag = src.get_min_max_mag()
prev_max_mag = self.max_mag
if prev_max_mag is None or max_mag > prev_max_mag:
self.max_mag = max_mag
[docs] def count_ruptures(self):
"""
Set src.num_ruptures on each source in the group
"""
for src in self:
src.nsites = 1
src.num_ruptures = src.count_ruptures()
print(src.weight)
return self
[docs] def split(self, maxweight):
"""
Split the group in subgroups with weight <= maxweight, unless it
it atomic.
"""
if self.atomic:
return [self]
out = []
for block in block_splitter(
self, maxweight, operator.attrgetter('weight')):
sg = copy.copy(self)
sg.sources = block
out.append(sg)
return out
[docs] def get_tom_toml(self, time_span):
"""
:returns: the TOM as a json string {'PoissonTOM': {'time_span': 50}}
"""
tom = self.temporal_occurrence_model
if tom is None:
return '[PoissonTOM]\ntime_span=%s' % time_span
dic = {tom.__class__.__name__: vars(tom)}
return toml.dumps(dic)
def __repr__(self):
return '<%s %s, %d source(s), weight=%d>' % (
self.__class__.__name__, self.trt, len(self.sources), self.weight)
def __lt__(self, other):
"""
Make sure there is a precise ordering of SourceGroup objects.
Objects with less sources are put first; in case the number
of sources is the same, use lexicographic ordering on the trts
"""
num_sources = len(self.sources)
other_sources = len(other.sources)
if num_sources == other_sources:
return self.trt < other.trt
return num_sources < other_sources
def __getitem__(self, i):
return self.sources[i]
def __iter__(self):
return iter(self.sources)
def __len__(self):
return len(self.sources)
def __toh5__(self):
lst = []
for i, src in enumerate(self.sources):
buf = pickle.dumps(src, pickle.HIGHEST_PROTOCOL)
lst.append((src.id, src.num_ruptures,
numpy.frombuffer(buf, numpy.uint8)))
attrs = dict(
trt=self.trt,
name=self.name or '',
src_interdep=self.src_interdep,
rup_interdep=self.rup_interdep,
grp_probability=self.grp_probability or '1')
return numpy.array(lst, source_dt), attrs
def __fromh5__(self, array, attrs):
vars(self).update(attrs)
self.sources = []
for row in array:
self.sources.append(pickle.loads(memoryview(row['pik'])))
[docs]def split_coords_2d(seq):
"""
:param seq: a flat list with lons and lats
:returns: a validated list of pairs (lon, lat)
>>> split_coords_2d([1.1, 2.1, 2.2, 2.3])
[(1.1, 2.1), (2.2, 2.3)]
"""
lons, lats = [], []
for i, el in enumerate(seq):
if i % 2 == 0:
lons.append(valid.longitude(el))
elif i % 2 == 1:
lats.append(valid.latitude(el))
return list(zip(lons, lats))
[docs]def split_coords_3d(seq):
"""
:param seq: a flat list with lons, lats and depths
:returns: a validated list of (lon, lat, depths) triplets
>>> split_coords_3d([1.1, 2.1, 0.1, 2.3, 2.4, 0.1])
[(1.1, 2.1, 0.1), (2.3, 2.4, 0.1)]
"""
lons, lats, depths = [], [], []
for i, el in enumerate(seq):
if i % 3 == 0:
lons.append(valid.longitude(el))
elif i % 3 == 1:
lats.append(valid.latitude(el))
elif i % 3 == 2:
depths.append(valid.depth(el))
return list(zip(lons, lats, depths))
[docs]def convert_nonParametricSeismicSource(fname, node, rup_spacing=5.0):
"""
Convert the given node into a non parametric source object.
:param fname:
full pathname to the XML file associated to the node
:param node:
a Node object coming from an XML file
:param rup_spacing:
Rupture spacing [km]
:returns:
a :class:`openquake.hazardlib.source.NonParametricSeismicSource`
instance
"""
trt = node.attrib.get('tectonicRegion')
rups_weights = None
if 'rup_weights' in node.attrib:
rups_weights = F64(node['rup_weights'].split())
nps = source.NonParametricSeismicSource(
node['id'], node['name'], trt, [], [])
nps.splittable = 'rup_weights' not in node.attrib
if fname:
path = os.path.splitext(fname)[0] + '.hdf5'
hdf5_fname = path if os.path.exists(path) else None
if hdf5_fname and node.text is None:
# gridded source, read the rupture data from the HDF5 file
with hdf5.File(hdf5_fname, 'r') as h:
dic = {k: d[:] for k, d in h[node['id']].items()}
nps.fromdict(dic, rups_weights)
return nps
# read the rupture data from the XML nodes
num_probs = None
for i, rupnode in enumerate(node):
po = rupnode['probs_occur']
probs = pmf.PMF(valid.pmf(po))
if num_probs is None: # first time
num_probs = len(probs.data)
elif len(probs.data) != num_probs:
# probs_occur must have uniform length for all ruptures
raise ValueError(
'prob_occurs=%s has %d elements, expected %s'
% (po, len(probs.data), num_probs))
rup = RuptureConverter(rup_spacing).convert_node(rupnode)
rup.tectonic_region_type = trt
rup.weight = None if rups_weights is None else rups_weights[i]
nps.data.append((rup, probs))
return nps
[docs]class RuptureConverter(object):
"""
Convert ruptures from nodes into Hazardlib ruptures.
"""
fname = None # should be set externally
def __init__(self, rupture_mesh_spacing, complex_fault_mesh_spacing=None):
self.rupture_mesh_spacing = rupture_mesh_spacing
self.complex_fault_mesh_spacing = (
complex_fault_mesh_spacing or rupture_mesh_spacing)
[docs] def get_mag_rake_hypo(self, node):
with context(self.fname, node):
mag = ~node.magnitude
rake = ~node.rake
h = node.hypocenter
hypocenter = geo.Point(h['lon'], h['lat'], h['depth'])
return mag, rake, hypocenter
[docs] def convert_node(self, node):
"""
Convert the given rupture node into a hazardlib rupture, depending
on the node tag.
:param node: a node representing a rupture
"""
return getattr(self, 'convert_' + striptag(node.tag))(node)
[docs] def geo_line(self, edge):
"""
Utility function to convert a node of kind edge
into a :class:`openquake.hazardlib.geo.Line` instance.
:param edge: a node describing an edge
"""
with context(self.fname, edge.LineString.posList) as plist:
coords = split_coords_2d(~plist)
return geo.Line([geo.Point(*p) for p in coords])
[docs] def geo_lines(self, edges):
"""
Utility function to convert a list of edges into a list of
:class:`openquake.hazardlib.geo.Line` instances.
:param edge: a node describing an edge
"""
lines = []
for edge in edges:
with context(self.fname, edge):
coords = split_coords_3d(~edge.LineString.posList)
lines.append(geo.Line([geo.Point(*p) for p in coords]))
return lines
[docs] def geo_planar(self, surface):
"""
Utility to convert a PlanarSurface node with subnodes
topLeft, topRight, bottomLeft, bottomRight into a
:class:`openquake.hazardlib.geo.PlanarSurface` instance.
:param surface: PlanarSurface node
"""
with context(self.fname, surface):
tl = surface.topLeft
top_left = geo.Point(tl['lon'], tl['lat'], tl['depth'])
tr = surface.topRight
top_right = geo.Point(tr['lon'], tr['lat'], tr['depth'])
bl = surface.bottomLeft
bottom_left = geo.Point(bl['lon'], bl['lat'], bl['depth'])
br = surface.bottomRight
bottom_right = geo.Point(br['lon'], br['lat'], br['depth'])
return geo.PlanarSurface.from_corner_points(
top_left, top_right, bottom_right, bottom_left)
[docs] def convert_surfaces(self, surface_nodes, sec_id=''):
"""
:param surface_nodes: surface nodes as described below
Utility to convert a list of surface nodes into a single hazardlib
surface. There are four possibilities:
1. there is a single simpleFaultGeometry node; returns a
:class:`openquake.hazardlib.geo.simpleFaultSurface` instance
2. there is a single complexFaultGeometry node; returns a
:class:`openquake.hazardlib.geo.complexFaultSurface` instance
3. there is a single griddedSurface node; returns a
:class:`openquake.hazardlib.geo.GriddedSurface` instance
4. there is either a single planarSurface or a list of planarSurface
nodes; returns a :class:`openquake.hazardlib.geo.PlanarSurface`
instance or a :class:`openquake.hazardlib.geo.MultiSurface` instance
5. there is either a single kiteSurface or a list of kiteSurface
nodes; returns a :class:`openquake.hazardlib.geo.KiteSurface`
instance or a :class:`openquake.hazardlib.geo.MultiSurface` instance
"""
surface_node = surface_nodes[0]
if surface_node.tag.endswith('simpleFaultGeometry'):
surface = geo.SimpleFaultSurface.from_fault_data(
self.geo_line(surface_node),
~surface_node.upperSeismoDepth,
~surface_node.lowerSeismoDepth,
~surface_node.dip,
self.rupture_mesh_spacing)
elif surface_node.tag.endswith('complexFaultGeometry'):
surface = geo.ComplexFaultSurface.from_fault_data(
self.geo_lines(surface_node),
self.complex_fault_mesh_spacing)
elif surface_node.tag.endswith('griddedSurface'):
with context(self.fname, surface_node):
coords = split_coords_3d(~surface_node.posList)
points = [geo.Point(*p) for p in coords]
surface = geo.GriddedSurface.from_points_list(points)
elif surface_node.tag.endswith('kiteSurface'):
# single or multiple kite surfaces
profs = []
for surface_node in surface_nodes:
profs.append(self.geo_lines(surface_node))
if len(profs) < 2:
surface = geo.KiteSurface.from_profiles(
profs[0], self.rupture_mesh_spacing,
self.rupture_mesh_spacing, sec_id=sec_id)
else:
surfaces = []
for prof in profs:
surfaces.append(geo.KiteSurface.from_profiles(
prof, self.rupture_mesh_spacing,
self.rupture_mesh_spacing))
surface = geo.MultiSurface(surfaces)
else: # a collection of planar surfaces
if len(surface_nodes) == 1:
return self.geo_planar(surface_nodes[0])
planar_surfaces = list(map(self.geo_planar, surface_nodes))
surface = geo.MultiSurface(planar_surfaces)
return surface
[docs] def convert_simpleFaultRupture(self, node):
"""
Convert a simpleFaultRupture node.
:param node: the rupture node
"""
mag, rake, hypocenter = self.get_mag_rake_hypo(node)
with context(self.fname, node):
surfaces = [node.simpleFaultGeometry]
rupt = source.rupture.BaseRupture(
mag=mag, rake=rake, tectonic_region_type=None,
hypocenter=hypocenter,
surface=self.convert_surfaces(surfaces))
return rupt
[docs] def convert_complexFaultRupture(self, node):
"""
Convert a complexFaultRupture node.
:param node: the rupture node
"""
mag, rake, hypocenter = self.get_mag_rake_hypo(node)
with context(self.fname, node):
[surface] = node.getnodes('complexFaultGeometry')
rupt = source.rupture.BaseRupture(
mag=mag, rake=rake, tectonic_region_type=None,
hypocenter=hypocenter,
surface=self.convert_surfaces([surface]))
return rupt
[docs] def convert_singlePlaneRupture(self, node):
"""
Convert a singlePlaneRupture node.
:param node: the rupture node
"""
mag, rake, hypocenter = self.get_mag_rake_hypo(node)
with context(self.fname, node):
surfaces = [node.planarSurface]
rupt = source.rupture.BaseRupture(
mag=mag, rake=rake,
tectonic_region_type=None,
hypocenter=hypocenter,
surface=self.convert_surfaces(surfaces))
return rupt
# used in scenarios or nonparametric sources
[docs] def convert_multiPlanesRupture(self, node):
"""
Convert a multiPlanesRupture node.
:param node: the rupture node
"""
mag, rake, hypocenter = self.get_mag_rake_hypo(node)
with context(self.fname, node):
if hasattr(node, 'planarSurface'):
surfaces = list(node.getnodes('planarSurface'))
for s in surfaces:
assert s.tag.endswith('planarSurface')
elif hasattr(node, 'kiteSurface'):
surfaces = list(node.getnodes('kiteSurface'))
for s in surfaces:
assert s.tag.endswith('kiteSurface')
else:
raise ValueError('Only multiSurfaces of planarSurfaces or'
'kiteSurfaces are supported (no mix)')
rupt = source.rupture.BaseRupture(
mag=mag, rake=rake,
tectonic_region_type=None,
hypocenter=hypocenter,
surface=self.convert_surfaces(surfaces))
return rupt
[docs] def convert_griddedRupture(self, node):
"""
Convert a griddedRupture node.
:param node: the rupture node
"""
mag, rake, hypocenter = self.get_mag_rake_hypo(node)
with context(self.fname, node):
surfaces = [node.griddedSurface]
rupt = source.rupture.BaseRupture(
mag=mag, rake=rake,
tectonic_region_type=None,
hypocenter=hypocenter,
surface=self.convert_surfaces(surfaces))
return rupt
[docs] def convert_ruptureCollection(self, node):
"""
:param node: a ruptureCollection node
:returns: a dictionary trt_smr -> EBRuptures
"""
coll = {}
for grpnode in node:
trt_smr = int(grpnode['id'])
coll[trt_smr] = ebrs = []
for node in grpnode:
rup = self.convert_node(node)
rup.rup_id = int(node['id'])
sesnodes = node.stochasticEventSets
n = 0 # number of events
for sesnode in sesnodes:
with context(self.fname, sesnode):
n += len(sesnode.text.split())
ebr = source.rupture.EBRupture(rup, 0, 0, numpy.array([n]))
ebrs.append(ebr)
return coll
[docs]class SourceConverter(RuptureConverter):
"""
Convert sources from valid nodes into Hazardlib objects.
"""
def __init__(self, investigation_time=50., rupture_mesh_spacing=5.,
complex_fault_mesh_spacing=None, width_of_mfd_bin=1.0,
area_source_discretization=None,
minimum_magnitude={'default': 0},
source_id=None, discard_trts=(),
floating_x_step=0, floating_y_step=0,
source_nodes=()):
self.investigation_time = investigation_time
self.area_source_discretization = area_source_discretization
self.minimum_magnitude = minimum_magnitude
self.rupture_mesh_spacing = rupture_mesh_spacing
self.complex_fault_mesh_spacing = (
complex_fault_mesh_spacing or rupture_mesh_spacing)
self.width_of_mfd_bin = width_of_mfd_bin
self.source_id = source_id
self.discard_trts = discard_trts
self.floating_x_step = floating_x_step
self.floating_y_step = floating_y_step
self.source_nodes = source_nodes
[docs] def convert_node(self, node):
"""
Convert the given source node into a hazardlib source, depending
on the node tag.
:param node: a node representing a source or a SourceGroup
"""
trt = node.attrib.get('tectonicRegion')
if trt and trt in self.discard_trts:
return
name = striptag(node.tag)
if name.endswith('Source'): # source node
source_id = node['id']
if self.source_id and source_id not in self.source_id:
# if source_id is set in the job.ini, discard all other sources
return
elif self.source_nodes and name not in self.source_nodes:
# if source_nodes is set, discard all other source nodes
return
obj = getattr(self, 'convert_' + name)(node)
if hasattr(obj, 'mfd') and hasattr(obj.mfd, 'slip_rate'):
# TruncatedGRMFD with slip rate (for Slovenia)
m = obj.mfd
obj.mfd = m.from_slip_rate(
m.min_mag, m.max_mag, m.bin_width, m.b_val,
m.slip_rate, m.rigidity, obj.get_fault_surface_area())
return obj
[docs] def convert_geometryModel(self, node):
"""
:param node: a geometryModel node
:returns: a dictionary sec_id -> section
"""
sections = {secnode["id"]: self.convert_node(secnode)
for secnode in node}
return sections
[docs] def convert_section(self, node):
"""
:param node: a section node
:returns: a list of surfaces
"""
with context(self.fname, node):
if hasattr(node, 'planarSurface'):
surfaces = list(node.getnodes('planarSurface'))
elif hasattr(node, 'kiteSurface'):
surfaces = list(node.getnodes('kiteSurface'))
else:
raise ValueError('Only planarSurfaces or kiteSurfaces ' +
'supported')
return self.convert_surfaces(surfaces, node['id'])
[docs] def get_tom(self, node):
"""
Convert the given node into a Temporal Occurrence Model object.
:param node: a node of kind poissonTOM or similar
:returns: a :class:`openquake.hazardlib.tom.BaseTOM` instance
"""
occurrence_rate = node.get('occurrence_rate')
kwargs = {}
# the occurrence_rate is not None only for clusters of sources,
# the ones implemented in calc.hazard_curve, see test case_35
if occurrence_rate:
tom_cls = tom.registry['ClusterPoissonTOM']
return tom_cls(self.investigation_time, occurrence_rate)
if 'tom' in node.attrib:
tom_cls = tom.registry[node['tom']]
# if tom is negbinom, sets mu and alpha attr to tom_class
if node['tom'] == 'NegativeBinomialTOM':
kwargs = {'alpha': float(node['alpha']),
'mu': float(node['mu'])}
else:
tom_cls = tom.registry['PoissonTOM']
return tom_cls(time_span=self.investigation_time, **kwargs)
[docs] def convert_mfdist(self, node):
"""
Convert the given node into a Magnitude-Frequency Distribution
object.
:param node: a node of kind incrementalMFD or truncGutenbergRichterMFD
:returns: a :class:`openquake.hazardlib.mfd.EvenlyDiscretizedMFD.` or
:class:`openquake.hazardlib.mfd.TruncatedGRMFD` instance
"""
with context(self.fname, node):
[mfd_node] = [subnode for subnode in node
if subnode.tag.endswith(KNOWN_MFDS)]
with context(self.fname, mfd_node):
if mfd_node.tag.endswith('incrementalMFD'):
return mfd.EvenlyDiscretizedMFD(
min_mag=mfd_node['minMag'], bin_width=mfd_node['binWidth'],
occurrence_rates=~mfd_node.occurRates)
elif mfd_node.tag.endswith('truncGutenbergRichterMFD'):
slip_rate = mfd_node.get('slipRate')
rigidity = mfd_node.get('rigidity')
if slip_rate:
assert rigidity
# instantiate with an area of 1, to be fixed later on
gr_mfd = mfd.TruncatedGRMFD.from_slip_rate(
mfd_node['minMag'], mfd_node['maxMag'],
self.width_of_mfd_bin, mfd_node['bValue'],
slip_rate, rigidity, area=1)
else:
gr_mfd = mfd.TruncatedGRMFD(
a_val=mfd_node['aValue'], b_val=mfd_node['bValue'],
min_mag=mfd_node['minMag'], max_mag=mfd_node['maxMag'],
bin_width=self.width_of_mfd_bin)
return gr_mfd
elif mfd_node.tag.endswith('arbitraryMFD'):
return mfd.ArbitraryMFD(
magnitudes=~mfd_node.magnitudes,
occurrence_rates=~mfd_node.occurRates)
elif mfd_node.tag.endswith('YoungsCoppersmithMFD'):
return mfd.YoungsCoppersmith1985MFD(
min_mag=mfd_node["minMag"],
b_val=mfd_node["bValue"],
char_mag=mfd_node["characteristicMag"],
char_rate=mfd_node.get("characteristicRate"),
total_moment_rate=mfd_node.get("totalMomentRate"),
bin_width=mfd_node["binWidth"])
elif mfd_node.tag.endswith('multiMFD'):
return mfd.multi_mfd.MultiMFD.from_node(
mfd_node, self.width_of_mfd_bin)
elif mfd_node.tag.endswith('taperedGutenbergRichterMFD'):
return mfd.TaperedGRMFD(
mfd_node['minMag'], mfd_node['maxMag'],
mfd_node['cornerMag'], self.width_of_mfd_bin,
mfd_node['aValue'], mfd_node['bValue'])
[docs] def convert_npdist(self, node):
"""
Convert the given node into a Nodal Plane Distribution.
:param node: a nodalPlaneDist node
:returns: a :class:`openquake.hazardlib.geo.NodalPlane` instance
"""
with context(self.fname, node):
npnode = node.nodalPlaneDist
npdist = []
for np in npnode:
prob, strike, dip, rake = (
np['probability'], np['strike'], np['dip'], np['rake'])
npdist.append((prob, geo.NodalPlane(strike, dip, rake)))
with context(self.fname, npnode):
fix_dupl(npdist, self.fname, npnode.lineno)
return pmf.PMF(npdist)
[docs] def convert_hddist(self, node):
"""
Convert the given node into a probability mass function for the
hypo depth distribution.
:param node: a hypoDepthDist node
:returns: a :class:`openquake.hazardlib.pmf.PMF` instance
"""
with context(self.fname, node):
hdnode = node.hypoDepthDist
hddist = [(hd['probability'], hd['depth']) for hd in hdnode]
with context(self.fname, hdnode):
fix_dupl(hddist, self.fname, hdnode.lineno)
return pmf.PMF(hddist)
[docs] def convert_areaSource(self, node):
"""
Convert the given node into an area source object.
:param node: a node with tag areaGeometry
:returns: a :class:`openquake.hazardlib.source.AreaSource` instance
"""
geom = node.areaGeometry
coords = split_coords_2d(~geom.Polygon.exterior.LinearRing.posList)
polygon = geo.Polygon([geo.Point(*xy) for xy in coords])
msr = ~node.magScaleRel
area_discretization = geom.attrib.get(
'discretization', self.area_source_discretization)
if area_discretization is None:
raise ValueError(
'The source %r has no `discretization` parameter and the job.'
'ini file has no `area_source_discretization` parameter either'
% node['id'])
return source.AreaSource(
source_id=node['id'],
name=node['name'],
tectonic_region_type=node.attrib.get('tectonicRegion'),
mfd=self.convert_mfdist(node),
rupture_mesh_spacing=self.rupture_mesh_spacing,
magnitude_scaling_relationship=msr,
rupture_aspect_ratio=~node.ruptAspectRatio,
upper_seismogenic_depth=~geom.upperSeismoDepth,
lower_seismogenic_depth=~geom.lowerSeismoDepth,
nodal_plane_distribution=self.convert_npdist(node),
hypocenter_distribution=self.convert_hddist(node),
polygon=polygon,
area_discretization=area_discretization,
temporal_occurrence_model=self.get_tom(node))
[docs] def convert_pointSource(self, node):
"""
Convert the given node into a point source object.
:param node: a node with tag pointGeometry
:returns: a :class:`openquake.hazardlib.source.PointSource` instance
"""
geom = node.pointGeometry
lon_lat = ~geom.Point.pos
msr = ~node.magScaleRel
return source.PointSource(
source_id=node['id'],
name=node['name'],
tectonic_region_type=node.attrib.get('tectonicRegion'),
mfd=self.convert_mfdist(node),
rupture_mesh_spacing=self.rupture_mesh_spacing,
magnitude_scaling_relationship=msr,
rupture_aspect_ratio=~node.ruptAspectRatio,
upper_seismogenic_depth=~geom.upperSeismoDepth,
lower_seismogenic_depth=~geom.lowerSeismoDepth,
location=geo.Point(*lon_lat),
nodal_plane_distribution=self.convert_npdist(node),
hypocenter_distribution=self.convert_hddist(node),
temporal_occurrence_model=self.get_tom(node))
[docs] def convert_multiPointSource(self, node):
"""
Convert the given node into a MultiPointSource object.
:param node: a node with tag multiPointGeometry
:returns: a :class:`openquake.hazardlib.source.MultiPointSource`
"""
geom = node.multiPointGeometry
lons, lats = zip(*split_coords_2d(~geom.posList))
msr = ~node.magScaleRel
return source.MultiPointSource(
source_id=node['id'],
name=node['name'],
tectonic_region_type=node.attrib.get('tectonicRegion'),
mfd=self.convert_mfdist(node),
magnitude_scaling_relationship=msr,
rupture_aspect_ratio=~node.ruptAspectRatio,
upper_seismogenic_depth=~geom.upperSeismoDepth,
lower_seismogenic_depth=~geom.lowerSeismoDepth,
nodal_plane_distribution=self.convert_npdist(node),
hypocenter_distribution=self.convert_hddist(node),
mesh=geo.Mesh(F32(lons), F32(lats)),
temporal_occurrence_model=self.get_tom(node))
[docs] def convert_simpleFaultSource(self, node):
"""
Convert the given node into a simple fault object.
:param node: a node with tag areaGeometry
:returns: a :class:`openquake.hazardlib.source.SimpleFaultSource`
instance
"""
geom = node.simpleFaultGeometry
msr = ~node.magScaleRel
fault_trace = self.geo_line(geom)
mfd = self.convert_mfdist(node)
with context(self.fname, node):
try:
hypo_list = valid.hypo_list(node.hypoList)
except AttributeError:
hypo_list = ()
try:
slip_list = valid.slip_list(node.slipList)
except AttributeError:
slip_list = ()
simple = source.SimpleFaultSource(
source_id=node['id'],
name=node['name'],
tectonic_region_type=node.attrib.get('tectonicRegion'),
mfd=mfd,
rupture_mesh_spacing=self.rupture_mesh_spacing,
magnitude_scaling_relationship=msr,
rupture_aspect_ratio=~node.ruptAspectRatio,
upper_seismogenic_depth=~geom.upperSeismoDepth,
lower_seismogenic_depth=~geom.lowerSeismoDepth,
fault_trace=fault_trace,
dip=~geom.dip,
rake=~node.rake,
temporal_occurrence_model=self.get_tom(node),
hypo_list=hypo_list,
slip_list=slip_list)
return simple
[docs] def convert_kiteFaultSource(self, node):
"""
Convert the given node into a kite fault object.
:param node: a node with tag kiteFaultSource
:returns: a :class:`openquake.hazardlib.source.KiteFaultSource`
instance
"""
as_kite = True
try:
geom = node.simpleFaultGeometry
fault_trace = self.geo_line(geom)
as_kite = False
except Exception:
geom = node.kiteSurface
profiles = self.geo_lines(geom)
msr = ~node.magScaleRel
mfd = self.convert_mfdist(node)
# get rupture floating steps
xstep = self.floating_x_step
ystep = self.floating_y_step
with context(self.fname, node):
if as_kite:
outsrc = source.KiteFaultSource(
source_id=node['id'],
name=node['name'],
tectonic_region_type=node.attrib.get('tectonicRegion'),
mfd=mfd,
rupture_mesh_spacing=self.rupture_mesh_spacing,
magnitude_scaling_relationship=msr,
rupture_aspect_ratio=~node.ruptAspectRatio,
temporal_occurrence_model=self.get_tom(node),
profiles=profiles,
floating_x_step=xstep,
floating_y_step=ystep,
rake=~node.rake,
profiles_sampling=None
)
else:
outsrc = source.KiteFaultSource.as_simple_fault(
source_id=node['id'],
name=node['name'],
tectonic_region_type=node.attrib.get('tectonicRegion'),
mfd=mfd,
rupture_mesh_spacing=self.rupture_mesh_spacing,
magnitude_scaling_relationship=msr,
rupture_aspect_ratio=~node.ruptAspectRatio,
upper_seismogenic_depth=~geom.upperSeismoDepth,
lower_seismogenic_depth=~geom.lowerSeismoDepth,
fault_trace=fault_trace,
dip=~geom.dip,
rake=~node.rake,
temporal_occurrence_model=self.get_tom(node),
floating_x_step=xstep,
floating_y_step=ystep)
return outsrc
[docs] def convert_complexFaultSource(self, node):
"""
Convert the given node into a complex fault object.
:param node: a node with tag areaGeometry
:returns: a :class:`openquake.hazardlib.source.ComplexFaultSource`
instance
"""
geom = node.complexFaultGeometry
edges = self.geo_lines(geom)
mfd = self.convert_mfdist(node)
msr = ~node.magScaleRel
with context(self.fname, node):
cmplx = source.ComplexFaultSource(
source_id=node['id'],
name=node['name'],
tectonic_region_type=node.attrib.get('tectonicRegion'),
mfd=mfd,
rupture_mesh_spacing=self.complex_fault_mesh_spacing,
magnitude_scaling_relationship=msr,
rupture_aspect_ratio=~node.ruptAspectRatio,
edges=edges,
rake=~node.rake,
temporal_occurrence_model=self.get_tom(node))
return cmplx
[docs] def convert_characteristicFaultSource(self, node):
"""
Convert the given node into a characteristic fault object.
:param node:
a characteristicFaultSource node
:returns:
a :class:`openquake.hazardlib.source.CharacteristicFaultSource`
instance
"""
char = source.CharacteristicFaultSource(
source_id=node['id'],
name=node['name'],
tectonic_region_type=node.attrib.get('tectonicRegion'),
mfd=self.convert_mfdist(node),
surface=self.convert_surfaces(node.surface),
rake=~node.rake,
temporal_occurrence_model=self.get_tom(node))
return char
[docs] def convert_nonParametricSeismicSource(self, node):
"""
Convert the given node into a non parametric source object.
:param node:
a node with tag areaGeometry
:returns:
a :class:`openquake.hazardlib.source.NonParametricSeismicSource`
instance
"""
return convert_nonParametricSeismicSource(
self.fname, node, self.rupture_mesh_spacing)
# used in UCERF
[docs] def convert_multiFaultSource(self, node):
"""
Convert the given node into a multi fault source object.
:param node:
a node with tag multiFaultSource
:returns:
a :class:`openquake.hazardlib.source.multiFaultSource`
instance
"""
sid = node.get('id')
name = node.get('name')
trt = node.get('tectonicRegion')
path = os.path.splitext(self.fname)[0] + '.hdf5'
hdf5_fname = path if os.path.exists(path) else None
if hdf5_fname and node.text is None:
# read the rupture data from the HDF5 file
with hdf5.File(hdf5_fname, 'r') as h:
dic = {k: d[:] for k, d in h[node['id']].items()}
with context(self.fname, node):
idxs = [x.decode('utf8').split() for x in dic['rupture_idxs']]
mags = rounded_unique(dic['mag'], idxs)
# NB: the sections will be fixed later on, in source_reader
mfs = MultiFaultSource(sid, name, trt, idxs, dic['probs_occur'],
dic['mag'], dic['rake'])
return mfs
probs = []
mags = []
rakes = []
idxs = []
num_probs = None
for i, rupnode in enumerate(node):
with context(self.fname, rupnode):
prb = valid.probabilities(rupnode['probs_occur'])
if num_probs is None: # first time
num_probs = len(prb)
elif len(prb) != num_probs:
# probs_occur must have uniform length for all ruptures
with context(self.fname, rupnode):
raise ValueError(
'prob_occurs=%s has %d elements, expected %s'
% (rupnode['probs_occur'], len(prb),
num_probs))
probs.append(prb)
mags.append(~rupnode.magnitude)
rakes.append(~rupnode.rake)
indexes = rupnode.sectionIndexes['indexes']
idxs.append(tuple(indexes.split(',')))
with context(self.fname, node):
mags = rounded_unique(mags, idxs)
rakes = numpy.array(rakes)
# NB: the sections will be fixed later on, in source_reader
mfs = MultiFaultSource(sid, name, trt, idxs, probs, mags, rakes)
return mfs
[docs] def convert_sourceModel(self, node):
return [self.convert_node(subnode) for subnode in node]
[docs] def convert_sourceGroup(self, node):
"""
Convert the given node into a SourceGroup object.
:param node:
a node with tag sourceGroup
:returns:
a :class:`SourceGroup` instance
"""
trt = node['tectonicRegion']
srcs_weights = node.attrib.get('srcs_weights')
grp_attrs = {k: v for k, v in node.attrib.items()
if k not in ('name', 'src_interdep', 'rup_interdep',
'srcs_weights')}
if node.attrib.get('src_interdep') != 'mutex':
# ignore weights set to 1 in old versions of the engine
srcs_weights = None
sg = SourceGroup(trt, min_mag=self.minimum_magnitude)
sg.temporal_occurrence_model = self.get_tom(node)
sg.name = node.attrib.get('name')
# Set attributes related to occurrence
sg.src_interdep = node.attrib.get('src_interdep', 'indep')
sg.rup_interdep = node.attrib.get('rup_interdep', 'indep')
sg.grp_probability = node.attrib.get('grp_probability', 1)
# Set the cluster attribute
sg.cluster = node.attrib.get('cluster') == 'true'
# Filter admitted cases
# 1. The source group is a cluster. In this case the cluster must have
# the attributes required to define its occurrence in time.
if sg.cluster:
msg = 'A cluster group requires the definition of a temporal'
msg += ' occurrence model'
assert 'tom' in node.attrib, msg
if isinstance(tom, PoissonTOM):
# hack in place of a ClusterPoissonTOM
assert hasattr(sg, 'occurrence_rate')
for src_node in node:
src = self.convert_node(src_node)
if src is None: # filtered out by source_id
continue
# transmit the group attributes to the underlying source
for attr, value in grp_attrs.items():
if attr == 'tectonicRegion':
src_trt = src_node.get('tectonicRegion')
if src_trt and src_trt != trt:
with context(self.fname, src_node):
raise ValueError('Found %s, expected %s' %
(src_node['tectonicRegion'], trt))
src.tectonic_region_type = trt
elif attr == 'grp_probability':
pass # do not transmit
else: # transmit as it is
setattr(src, attr, node[attr])
sg.update(src)
if sg.src_interdep == 'mutex':
if len(node) and len(srcs_weights) != len(node):
raise ValueError(
'There are %d srcs_weights but %d source(s) in %s'
% (len(srcs_weights), len(node), self.fname))
tot = 0
for src, sw in zip(sg, srcs_weights):
src.mutex_weight = sw
tot += sw
with context(self.fname, node):
numpy.testing.assert_allclose(
tot, 1., err_msg='sum(srcs_weights)', atol=5E-6)
# check that, when the cluster option is set, the group has a temporal
# occurrence model properly defined
if sg.cluster and not hasattr(sg, 'temporal_occurrence_model'):
msg = 'The Source Group is a cluster but does not have a '
msg += 'temporal occurrence model'
raise ValueError(msg)
return sg
[docs]@dataclass
class Row:
id: str
name: str
code: str
groupname: str
tectonicregion: str
mfd: str
magscalerel: str
ruptaspectratio: float
upperseismodepth: float
lowerseismodepth: float
nodalplanedist: list
hypodepthdist: list
hypoList: list
slipList: list
rake: float
geomprops: list
geom: str
coords: list
wkt: str
Row.__init__.__defaults__ = ('',) # wkt
[docs]@dataclass
class NPRow:
id: str
name: str
code: str
tectonicregion: str
geom: str
coords: list
wkt: str
def _planar(surface):
poly = []
tl = surface.topLeft
poly.append((tl['lon'], tl['lat'], tl['depth']))
tr = surface.topRight
poly.append((tr['lon'], tr['lat'], tr['depth']))
br = surface.bottomRight
poly.append((br['lon'], br['lat'], br['depth']))
bl = surface.bottomLeft
poly.append((bl['lon'], bl['lat'], bl['depth']))
poly.append((tl['lon'], tl['lat'], tl['depth'])) # close the polygon
return [poly]
[docs]class RowConverter(SourceConverter):
"""
Used in the command oq nrml_to_csv to convert source models into
Row objects.
"""
[docs] def convert_node(self, node):
"""
Convert the given source node into a Row object
"""
trt = node.attrib.get('tectonicRegion')
if trt and trt in self.discard_trts:
return
with context(self.fname, node):
return getattr(self, 'convert_' + striptag(node.tag))(node)
[docs] def convert_mfdist(self, node):
with context(self.fname, node):
[mfd_node] = [subnode for subnode in node
if subnode.tag.endswith(KNOWN_MFDS)]
return str(node_to_dict(mfd_node))
[docs] def convert_npdist(self, node):
lst = []
for w, np in super().convert_npdist(node).data:
dic = {'probability': w, 'dip': np.dip, 'rake': np.rake,
'strike': np.strike}
lst.append(dic)
return str(lst)
[docs] def convert_hddist(self, node):
lst = []
for w, hd in super().convert_hddist(node).data:
lst.append(dict(probability=w, depth=hd))
return str(lst)
[docs] def convert_hypolist(self, node):
try:
hypo_list = node.hypoList
except AttributeError:
lst = []
else:
lst = [{'alongStrike': hl['alongStrike'],
'downDip': hl['downDip'],
'weight': hl['weight']} for hl in hypo_list]
return str(lst)
[docs] def convert_sliplist(self, node):
try:
slip_list = node.slipList
except AttributeError:
lst = []
else:
lst = [node_to_dict(n)['slip'] for n in slip_list.nodes]
return str(lst)
[docs] def convert_rake(self, node):
try:
return ~node.rake
except AttributeError:
return ''
[docs] def convert_geomprops(self, node):
# NOTE: node_to_dict(node) returns a dict having the geometry type as
# key and the corresponding properties as value, so we get the first
# value to retrieve the information we need
full_geom_props = list(node_to_dict(node).values())[0]
geom_props = {k: full_geom_props[k] for k in full_geom_props
if k not in EXCLUDE_FROM_GEOM_PROPS}
return str(geom_props)
[docs] def convert_areaSource(self, node):
geom = node.areaGeometry
coords = split_coords_2d(~geom.Polygon.exterior.LinearRing.posList)
if coords[0] != coords[-1]:
coords += [coords[0]] # close the polygon
return Row(
node['id'],
node['name'],
'A',
node.get('groupname', ''),
node.get('tectonicRegion', ''),
self.convert_mfdist(node),
str(~node.magScaleRel),
~node.ruptAspectRatio,
~geom.upperSeismoDepth,
~geom.lowerSeismoDepth,
self.convert_npdist(node),
self.convert_hddist(node),
self.convert_hypolist(node),
self.convert_sliplist(node),
self.convert_rake(node),
self.convert_geomprops(geom),
'Polygon', [coords])
[docs] def convert_pointSource(self, node):
geom = node.pointGeometry
return Row(
node['id'],
node['name'],
'P',
node.get('groupname', ''),
node.get('tectonicRegion', ''),
self.convert_mfdist(node),
str(~node.magScaleRel),
~node.ruptAspectRatio,
~geom.upperSeismoDepth,
~geom.lowerSeismoDepth,
self.convert_npdist(node),
self.convert_hddist(node),
self.convert_hypolist(node),
self.convert_sliplist(node),
self.convert_rake(node),
self.convert_geomprops(geom),
'Point', ~geom.Point.pos)
[docs] def convert_multiPointSource(self, node):
geom = node.multiPointGeometry
coords = split_coords_2d(~geom.posList)
return Row(
node['id'],
node['name'],
'M',
node.get('groupname', ''),
node.get('tectonicRegion', ''),
self.convert_mfdist(node),
str(~node.magScaleRel),
~node.ruptAspectRatio,
~geom.upperSeismoDepth,
~geom.lowerSeismoDepth,
self.convert_npdist(node),
self.convert_hddist(node),
self.convert_hypolist(node),
self.convert_sliplist(node),
self.convert_rake(node),
self.convert_geomprops(geom),
'MultiPoint', coords)
[docs] def convert_simpleFaultSource(self, node):
geom = node.simpleFaultGeometry
return Row(
node['id'],
node['name'],
'S',
node.get('groupname', ''),
node.get('tectonicRegion', ''),
self.convert_mfdist(node),
str(~node.magScaleRel),
~node.ruptAspectRatio,
~geom.upperSeismoDepth,
~geom.lowerSeismoDepth,
[],
[],
self.convert_hypolist(node),
self.convert_sliplist(node),
self.convert_rake(node),
self.convert_geomprops(geom),
'LineString', [(p.x, p.y) for p in self.geo_line(geom)])
[docs] def convert_complexFaultSource(self, node):
geom = node.complexFaultGeometry # 1005
edges = []
for line in self.geo_lines(geom):
edges.append([(p.x, p.y, p.z) for p in line])
return Row(
node['id'],
node['name'],
'C',
node.get('groupname', ''),
node.get('tectonicRegion', ''),
self.convert_mfdist(node),
str(~node.magScaleRel),
~node.ruptAspectRatio,
numpy.nan,
numpy.nan,
[],
[],
self.convert_hypolist(node),
self.convert_sliplist(node),
self.convert_rake(node),
self.convert_geomprops(geom),
'3D MultiLineString', edges)
[docs] def convert_characteristicFaultSource(self, node):
_, kind = node.surface[0].tag.split('}')
if kind == 'simpleFaultGeometry':
geom = 'LineString'
coords = [(point.x, point.y) for point in self.geo_line(
node.surface.simpleFaultGeometry)]
elif kind == 'complexFaultGeometry':
geom = '3D MultiLineString'
coords = []
for line in self.geo_lines(node.surface.complexFaultGeometry):
coords.append([(p.x, p.y, p.z) for p in line])
elif kind == 'planarSurface':
geom = '3D MultiPolygon'
coords = [_planar(surface) for surface in node.surface]
return Row(
node['id'],
node['name'],
'X',
node.get('groupname', ''),
node.get('tectonicRegion', ''),
self.convert_mfdist(node),
numpy.nan,
numpy.nan,
numpy.nan,
numpy.nan,
[{'rake': ~node.rake}],
[],
self.convert_hypolist(node),
self.convert_sliplist(node),
self.convert_rake(node),
self.convert_geomprops(node.surface),
geom, coords, '')
[docs] def convert_nonParametricSeismicSource(self, node):
nps = convert_nonParametricSeismicSource(self.fname, node)
return NPRow(
node['id'],
node['name'],
'N',
node['tectonicRegion'],
'Polygon', [nps.polygon.coords], '')
[docs] def convert_multiFaultSource(self, node):
mfs = super().convert_multiFaultSource(node)
return NPRow(node['id'], node['name'], 'F',
node.get('tectonicRegion', ''), 'Polygon', mfs, '')
# ################### MultiPointSource conversion ######################## #
[docs]def multikey(node):
"""
:returns: (usd, lsd, rar, hddist, npdist, magScaleRel) for the given node
"""
hd = tuple((node['probability'], node['depth'])
for node in node.hypoDepthDist)
npd = tuple(
((node['probability'], node['rake'], node['strike'], node['dip']))
for node in node.nodalPlaneDist)
geom = node.pointGeometry
return (round(~geom.upperSeismoDepth, 1), round(~geom.lowerSeismoDepth, 1),
~node.ruptAspectRatio, hd, npd, str(~node.magScaleRel))
[docs]def collapse(array):
"""
Collapse a homogeneous array into a scalar; do nothing if the array
is not homogenous
"""
if len(set(a for a in array)) == 1: # homogenous array
return array[0]
return array
[docs]def mfds2multimfd(mfds):
"""
Convert a list of MFD nodes into a single MultiMFD node
"""
_, kind = mfds[0].tag.split('}')
node = Node('multiMFD', dict(kind=kind, size=len(mfds)))
lengths = None
for field in mfd.multi_mfd.ASSOC[kind][1:]:
alias = mfd.multi_mfd.ALIAS.get(field, field)
if field in ('magnitudes', 'occurRates'):
data = [~getattr(m, field) for m in mfds]
lengths = [len(d) for d in data]
data = sum(data, []) # list of lists
else:
try:
data = [m[alias] for m in mfds]
except KeyError:
if alias == 'binWidth':
# missing bindWidth in GR MDFs is ok
continue
else:
raise
node.append(Node(field, text=collapse(data)))
if lengths: # this is the last field if present
node.append(Node('lengths', text=collapse(lengths)))
return node
def _pointsources2multipoints(srcs, i):
# converts pointSources with the same hddist, npdist and msr into a
# single multiPointSource.
allsources = []
for (usd, lsd, rar, hd, npd, msr), sources in groupby(
srcs, multikey).items():
if len(sources) == 1: # there is a single source
allsources.extend(sources)
continue
mfds = [src[3] for src in sources]
points = []
for src in sources:
pg = src.pointGeometry
points.extend(~pg.Point.pos)
geom = Node('multiPointGeometry')
geom.append(Node('gml:posList', text=points))
geom.append(Node('upperSeismoDepth', text=usd))
geom.append(Node('lowerSeismoDepth', text=lsd))
node = Node(
'multiPointSource',
dict(id='mps-%d' % i, name='multiPointSource-%d' % i),
nodes=[geom])
node.append(Node("magScaleRel", text=collapse(msr)))
node.append(Node("ruptAspectRatio", text=rar))
node.append(mfds2multimfd(mfds))
node.append(Node('nodalPlaneDist', nodes=[
Node('nodalPlane', dict(probability=prob, rake=rake,
strike=strike, dip=dip))
for prob, rake, strike, dip in npd]))
node.append(Node('hypoDepthDist', nodes=[
Node('hypoDepth', dict(depth=depth, probability=prob))
for prob, depth in hd]))
allsources.append(node)
i += 1
return i, allsources
[docs]def drop_trivial_weights(group):
ws = group.attrib.get('srcs_weights')
if ws and len(set(ws)) == 1: # all equal
del group.attrib['srcs_weights']
[docs]def update_source_model(sm_node, fname):
"""
:param sm_node: a sourceModel Node object containing sourceGroups
"""
i = 0
for group in sm_node:
if not group.tag.endswith('sourceGroup'):
raise InvalidFile('wrong NRML, got %s instead of '
'sourceGroup in %s' % (group.tag, fname))
drop_trivial_weights(group)
psrcs = []
others = []
for src in group:
try:
del src.attrib['tectonicRegion'] # make the trt implicit
except KeyError:
pass # already missing
if src.tag.endswith('pointSource'):
psrcs.append(src)
else:
others.append(src)
others.sort(key=lambda src: (src.tag, src['id']))
i, sources = _pointsources2multipoints(psrcs, i)
group.nodes = sources + others