Source code for weaver.processes.execution

import logging
import os
from time import sleep
from typing import TYPE_CHECKING

import colander
import psutil
from celery.exceptions import TimeoutError as CeleryTaskTimeoutError
from celery.utils.debug import ps as get_celery_process
from celery.utils.log import get_task_logger
from owslib.util import clean_ows_url
from owslib.wps import BoundingBoxDataInput, ComplexDataInput
from pyramid.httpexceptions import HTTPBadRequest, HTTPNotAcceptable
from pyramid_celery import celery_app as app

from weaver.database import get_db
from weaver.datatype import Process, Service
from weaver.execute import ExecuteControlOption, ExecuteMode
from weaver.formats import AcceptLanguage, ContentType, clean_media_type_format, repr_json
from weaver.notify import map_job_subscribers, notify_job_subscribers
from weaver.owsexceptions import OWSInvalidParameterValue, OWSNoApplicableCode
from weaver.processes import wps_package
from weaver.processes.constants import WPS_BOUNDINGBOX_DATA, WPS_COMPLEX_DATA, JobInputsOutputsSchema
from weaver.processes.convert import (
from weaver.processes.types import ProcessType
from weaver.status import JOB_STATUS_CATEGORIES, Status, StatusCategory, map_status
from import StoreJobs, StoreProcesses
from weaver.utils import (
from weaver.visibility import Visibility
from weaver.wps.utils import (
from weaver.wps_restapi import swagger_definitions as sd
from import get_job_results_response, get_job_submission_response
from weaver.wps_restapi.processes.utils import resolve_process_tag

[docs] LOGGER = logging.getLogger(__name__)
if TYPE_CHECKING: from typing import Dict, List, Optional, Tuple, Union from uuid import UUID from import Task from pyramid.request import Request from pywps.inout.inputs import BoundingBoxInput, ComplexInput from weaver.datatype import Job from weaver.processes.convert import OWS_Input_Type, ProcessOWS from weaver.status import StatusType from weaver.typedefs import ( AnyAcceptLanguageHeader, AnyProcessRef, AnyResponseType, AnyServiceRef, AnyValueType, CeleryResult, HeaderCookiesType, HeadersType, JobValueBbox, JSON, ProcessExecution, SettingsType, Statistics ) from weaver.visibility import AnyVisibility
[docs] class JobProgress(object): """ Job process execution progress. """
[docs] SETUP = 1
[docs] DESCRIBE = 2
[docs] GET_INPUTS = 3
[docs] GET_OUTPUTS = 4
[docs] NOTIFY = 99
[docs] DONE = 100
[docs] def execute_process(task, job_id, wps_url, headers=None): # type: (Task, UUID, str, Optional[HeadersType]) -> StatusType """ Celery task that executes the WPS process job monitoring as status updates (local and remote). """ from weaver.wps.service import get_pywps_service LOGGER.debug("Job execute process called.") task_process = get_celery_process() rss_start = task_process.memory_info().rss registry = get_registry(app) # local thread, whether locally or dispatched celery settings = get_settings(registry) db = get_db(registry, reset_connection=True) # reset the connection because we are in a forked celery process store = db.get_store(StoreJobs) job = store.fetch_by_id(job_id) job.started = now() job.status = Status.STARTED # will be mapped to 'RUNNING' job.status_message = f"Job {Status.STARTED}." # will preserve detail of STARTED vs RUNNING job.save_log(message=job.status_message) task_logger = get_task_logger(__name__) notify_job_subscribers(job, task_logger, settings) job.save_log(logger=task_logger, message="Job task setup initiated.") load_pywps_config(settings) job.progress = JobProgress.SETUP job.task_id = job.save_log(logger=task_logger, message="Job task setup completed.") job = store.update_job(job) # Flag to keep track if job is running in background (remote-WPS, CWL app, etc.). # If terminate signal is sent to worker task via API dismiss request while still running in background, # the raised exception within the task will switch the job to Status.FAILED, but this will not raise an # exception here. Since the task execution 'succeeds' without raising, it skips directly to the last 'finally'. # Patch it back to Status.DISMISSED in this case. task_terminated = True try: job.progress = JobProgress.DESCRIBE job.save_log(logger=task_logger, message=f"Employed WPS URL: [{wps_url!s}]", level=logging.DEBUG) job.save_log(logger=task_logger, message=f"Execute WPS request for process [{job.process!s}]") wps_process = fetch_wps_process(job, wps_url, headers, settings) # prepare inputs job.progress = JobProgress.GET_INPUTS job.save_log(logger=task_logger, message="Fetching job input definitions.") wps_inputs = parse_wps_inputs(wps_process, job) # prepare outputs job.progress = JobProgress.GET_OUTPUTS job.save_log(logger=task_logger, message="Fetching job output definitions.") wps_outputs = [(o.identifier, o.dataType == WPS_COMPLEX_DATA) for o in wps_process.processOutputs] # if process refers to a remote WPS provider, pass it down to avoid unnecessary re-fetch request if job.is_local: process = None # already got all the information needed pre-loaded in PyWPS service else: service = Service(name=job.service, url=wps_url) process = Process.from_ows(wps_process, service, settings) job.progress = JobProgress.EXECUTE_REQUEST job.save_log(logger=task_logger, message="Starting job process execution.") job.save_log(logger=task_logger, message="Following updates could take a while until the Application Package answers...") wps_worker = get_pywps_service(environ=settings, is_worker=True) execution = wps_worker.execute_job(job, wps_inputs=wps_inputs, wps_outputs=wps_outputs, remote_process=process, headers=headers) if not execution.process and execution.errors: raise execution.errors[0] # adjust status location wps_status_path = get_wps_local_status_location(execution.statusLocation, settings) job.progress = JobProgress.EXECUTE_STATUS_LOCATION LOGGER.debug("WPS status location that will be queried: [%s]", wps_status_path) if not wps_status_path.startswith("http") and not os.path.isfile(wps_status_path): LOGGER.warning("WPS status location not resolved to local path: [%s]", wps_status_path) job.save_log(logger=task_logger, level=logging.DEBUG, message=f"Updated job status location: [{wps_status_path}].") job.status = Status.RUNNING job.status_message = execution.statusMessage or f"{job!s} initiation done." job.status_location = wps_status_path job.request = execution.request job.response = execution.response job.progress = JobProgress.EXECUTE_MONITOR_START job.save_log(logger=task_logger, message="Starting monitoring of job execution.") job = store.update_job(job) max_retries = 5 num_retries = 0 run_step = 0 while execution.isNotComplete() or run_step == 0: if num_retries >= max_retries: job.save_log(errors=execution.errors, logger=task_logger) job = store.update_job(job) raise Exception(f"Could not read status document after {max_retries} retries. Giving up.") try: # NOTE: # Don't actually log anything here until process is completed (success or fail) so that underlying # WPS execution logs can be inserted within the current job log and appear continuously. # Only update internal job fields in case they get referenced elsewhere. progress_min = JobProgress.EXECUTE_MONITOR_LOOP progress_max = JobProgress.EXECUTE_MONITOR_DONE job.progress = progress_min run_delay = wait_secs(run_step) execution = check_wps_status(location=wps_status_path, settings=settings, sleep_secs=run_delay) job_msg = (execution.statusMessage or "").strip() job.response = execution.response job.status = map_status(execution.getStatus()) job_status_msg = job_msg or "n/a" job_percent = execution.percentCompleted job.status_message = f"Job execution monitoring (progress: {job_percent}%, status: {job_status_msg})." if execution.isComplete(): msg_progress = f" (status: {job_msg})" if job_msg else "" if execution.isSucceded(): wps_package.retrieve_package_job_log(execution, job, progress_min, progress_max) job.status = map_status(Status.SUCCEEDED) job.status_message = f"Job succeeded{msg_progress}." job.progress = progress_max job.save_log(logger=task_logger) job_results = [ ows2json_output_data(output, process, settings) for output in execution.processOutputs ] job.results = make_results_relative(job_results, settings) else: task_logger.debug("Job failed.") wps_package.retrieve_package_job_log(execution, job, progress_min, progress_max) job.status_message = f"Job failed{msg_progress}." job.progress = progress_max job.save_log(errors=execution.errors, logger=task_logger) task_logger.debug("Mapping Job references with generated WPS locations.") map_locations(job, settings) job = store.update_job(job) except Exception as exc: num_retries += 1 task_logger.debug("Exception raised: %s", repr(exc)) job.status_message = f"Could not read status XML document for {job!s}. Trying again..." job.save_log(errors=execution.errors, logger=task_logger) job = store.update_job(job) sleep(1) else: num_retries = 0 run_step += 1 finally: task_terminated = False # reached only if WPS execution completed (worker not terminated beforehand) job = store.update_job(job) except Exception as exc: # if 'execute_job' finishes quickly before even reaching the 'monitoring loop' # consider WPS execution produced an error (therefore Celery worker not terminated) task_terminated = False LOGGER.exception("Failed running [%s]", job) LOGGER.debug("Failed job [%s] raised an exception.", job, exc_info=exc) # note: don't update the progress here to preserve last one that was set job.status = map_status(Status.FAILED) job.status_message = f"Failed to run {job!s}." errors = f"{fully_qualified_name(exc)}: {exc!s}" job.save_log(errors=errors, logger=task_logger) job = store.update_job(job) finally: # note: # don't update the progress and status here except for 'success' to preserve last error that was set # it is more relevant to return the latest step that worked properly to understand where it failed job = store.fetch_by_id( # if task worker terminated, local 'job' is out of date compared to remote/background runner last update if task_terminated and map_status(job.status) == Status.FAILED: job.status = Status.DISMISSED task_success = map_status(job.status) not in JOB_STATUS_CATEGORIES[StatusCategory.FAILED] collect_statistics(task_process, settings, job, rss_start) if task_success: job.progress = JobProgress.EXECUTE_MONITOR_END job.status_message = f"Job {job.status}." job.save_log(logger=task_logger) if task_success: job.progress = JobProgress.NOTIFY notify_job_subscribers(job, task_logger, settings) if job.status not in JOB_STATUS_CATEGORIES[StatusCategory.FINISHED]: job.status = Status.SUCCEEDED job.status_message = f"Job {job.status}." job.mark_finished() if task_success: job.progress = JobProgress.DONE job.save_log(logger=task_logger, message="Job task complete.") job = store.update_job(job) return job.status
[docs] def collect_statistics(process, settings=None, job=None, rss_start=None): # type: (Optional[psutil.Process], Optional[SettingsType], Optional[Job], Optional[int]) -> Optional[Statistics] """ Collect any available execution statistics and store them in the :term:`Job` if provided. """ try: mem_used = None if job: mem_info = list(filter(lambda line: "cwltool" in line and "memory used" in line, job.logs)) mem_used = None if mem_info: mem_info = mem_info[0].split(":")[-1].strip() mem_used = parse_number_with_unit(mem_info, binary=True) stats = {} # type: JSON if mem_used: stats["application"] = { # see: 'cwltool.job.JobBase.process_monitor', reported memory in logs uses 'rss' "usedMemory": apply_number_with_unit(mem_used, binary=True), "usedMemoryBytes": mem_used, } rss = None if process: proc_info = process.memory_full_info() rss = getattr(proc_info, "rss", 0) uss = getattr(proc_info, "uss", 0) vms = getattr(proc_info, "vms", 0) stats["process"] = { "rss": apply_number_with_unit(rss, binary=True), "rssBytes": rss, "uss": apply_number_with_unit(uss, binary=True), "ussBytes": uss, "vms": apply_number_with_unit(vms, binary=True), "vmsBytes": vms, } fields = [("usedThreads", "num_threads"), ("usedCPU", "cpu_num"), ("usedHandles", "num_handles")] for field, method in fields: func = getattr(process, method, None) stats["process"][field] = func() if func is not None else 0 if rss_start and rss: # diff of RSS between start/end to consider only execution of the job steps # this more accurately reports used memory by the execution itself, omitting celery worker's base memory rss_diff = rss - rss_start stats["process"]["usedMemory"] = apply_number_with_unit(rss_diff, binary=True) stats["process"]["usedMemoryBytes"] = rss_diff total_size = 0 if job: stats["outputs"] = {} for result in job.results: res_ref = get_any_value(result, file=True) if res_ref and isinstance(res_ref, str): if res_ref.startswith(f"/{}"): # pseudo-relative reference out_dir = get_wps_output_dir(settings) res_ref = os.path.join(out_dir, res_ref.lstrip("/")) if os.path.isfile(res_ref): res_stat = os.stat(res_ref) res_id = get_any_id(result) res_size = res_stat.st_size stats["outputs"][res_id] = { "size": apply_number_with_unit(res_size, binary=True), "sizeBytes": res_size, } total_size += res_size stats["process"]["totalSize"] = apply_number_with_unit(total_size, binary=True) stats["process"]["totalSizeBytes"] = total_size if stats and job: job.statistics = stats return stats or None except Exception as exc: # pragma: no cover LOGGER.warning("Ignoring error that occurred during statistics collection [%s]", str(exc), exc_info=exc)
[docs] def fetch_wps_process(job, wps_url, headers, settings): # type: (Job, str, HeadersType, SettingsType) -> ProcessOWS """ Retrieves the WPS process description from the local or remote WPS reference URL. """ try: wps = get_wps_client(wps_url, settings, headers=headers, language=job.accept_language) raise_on_xml_exception(wps._capabilities) # noqa: W0212 except Exception as ex: job.save_log(errors=ex, message=f"Failed WPS client creation for process [{job.process!s}]") raise OWSNoApplicableCode(f"Failed to retrieve WPS capabilities. Error: [{ex!s}].") try: wps_process = wps.describeprocess(job.process) except Exception as ex: # pragma: no cover raise OWSNoApplicableCode(f"Failed to retrieve WPS process description. Error: [{ex!s}].") return wps_process
[docs] def parse_wps_input_format(input_info, type_field="mime_type", search_variations=True): # type: (JSON, str, bool) -> Tuple[Optional[str], Optional[str]] ctype = get_field(input_info, type_field, search_variations=search_variations, default=None) c_enc = get_field(input_info, "encoding", search_variations=True, default=None) if not c_enc: ctype_params = parse_kvp(ctype) c_enc = ctype_params.get("charset") c_enc = c_enc[0] if c_enc and isinstance(c_enc, list) else None return ctype, c_enc
[docs] def parse_wps_input_complex(input_value, input_info): # type: (Union[str, JSON], JSON) -> ComplexDataInput """ Parse the input data details into a complex input. """ # if provided, pass down specified input format to allow validation against supported formats c_enc = ctype = schema = None schema_vars = ["reference", "$schema"] input_field = get_any_value(input_info, key=True) if isinstance(input_value, dict): ctype, c_enc = parse_wps_input_format(input_value, "type", search_variations=False) if not ctype: ctype, c_enc = parse_wps_input_format(input_value) schema = get_field(input_value, "schema", search_variations=True, default=None, extra_variations=schema_vars) input_value = input_value[input_field] input_value = repr_json(input_value, indent=None, ensure_ascii=(c_enc in ["ASCII", "ascii"])) if not ctype: ctype, c_enc = parse_wps_input_format(input_info) media_format = get_field(input_info, "format", default=None) if not ctype and isinstance(media_format, dict): ctype, c_enc = parse_wps_input_format(media_format) if isinstance(schema, dict): schema = get_field(schema, "$ref", default=None, extra_variations=schema_vars) # need to support 'file://' scheme which could be omitted # to ensure owslib parses it has a link (asReference), add it if missing if input_field in ["href", "reference"] and "://" not in str(input_value): input_value = f"file://{input_value}" return ComplexDataInput(input_value, mimeType=ctype, encoding=c_enc, schema=schema)
[docs] def parse_wps_input_bbox(input_value, input_info): # type: (Union[str, JobValueBbox], JSON) -> BoundingBoxDataInput """ Parse the input data details into a bounding box input. """ bbox_crs = None bbox_val = input_value if isinstance(input_value, dict): bbox_crs = input_value.get("crs") bbox_val = input_value.get("bbox") if not bbox_crs: bbox_crs_def = input_info.get("bbox", {}) if isinstance(bbox_crs_def, dict) and "default" in bbox_crs_def: bbox_crs = bbox_crs_def["default"] or None bbox_val = bbox_val.split(",") if isinstance(bbox_val, str) else bbox_val bbox_dim = len(bbox_val) // 2 return BoundingBoxDataInput(bbox_val, crs=bbox_crs, dimensions=bbox_dim)
[docs] def parse_wps_input_literal(input_value): # type: (Union[AnyValueType, JSON]) -> Optional[str] """ Parse the input data details into a literal input. """ # if JSON 'null' was given, the execution content should simply omit the optional input # cannot distinguish directly between empty string and 'null' in XML representation if input_value is None: return None # measurement structure # however, owslib does not care about the UoM specified as input (no way to provide it) if isinstance(input_value, dict): val = get_any_value(input_value, file=False, default=input_value) # in case it was nested twice under 'value' if isinstance(val, dict): val = get_field(val, "measure", search_variations=True, default=val) if val is not None: input_value = val # need to use literal string for any data type return str(input_value)
[docs] def parse_wps_inputs(wps_process, job): # type: (ProcessOWS, Job) -> List[Tuple[str, OWS_Input_Type]] """ Parses expected WPS process inputs against submitted job input values considering supported process definitions. """ complex_inputs = {} # type: Dict[str, ComplexInput] bbox_inputs = {} # type: Dict[str, BoundingBoxInput] for process_input in wps_process.dataInputs: if process_input.dataType == WPS_COMPLEX_DATA: complex_inputs[process_input.identifier] = process_input elif process_input.dataType == WPS_BOUNDINGBOX_DATA: bbox_inputs[process_input.identifier] = process_input try: wps_inputs = [] # parse both dict and list type inputs job_inputs = job.inputs.items() if isinstance(job.inputs, dict) else job.get("inputs", []) for job_input in job_inputs: if isinstance(job_input, tuple): input_id = job_input[0] input_val = job_input[1] job_input = input_val else: input_id = get_any_id(job_input) input_val = get_any_value(job_input) if input_id in bbox_inputs and input_val is None: # inline bbox input_val = job_input # FIXME: handle minOccurs>=2 vs single-value inputs # - # - # in case of array inputs, must repeat (id, value) if isinstance(input_val, list): input_values = input_val input_details = input_val # each value has its own metadata else: input_values = [input_val] input_details = [job_input] # metadata directly in definition, not nested per array value for input_value, input_info in zip(input_values, input_details): if input_id in complex_inputs: input_data = parse_wps_input_complex(input_value, input_info) elif input_id in bbox_inputs: input_data = parse_wps_input_bbox(input_value, input_info) else: input_data = parse_wps_input_literal(input_value) if input_data is None: job.save_log( message=f"Removing [{input_id}] data input from execution request, value was 'null'.", logger=LOGGER, level=logging.WARNING, ) else: wps_inputs.append((input_id, input_data)) except KeyError: wps_inputs = [] return wps_inputs
[docs] def make_results_relative(results, settings): # type: (List[JSON], SettingsType) -> List[JSON] """ Converts file references to a pseudo-relative location to allow the application to dynamically generate paths. Redefines job results to be saved in database as pseudo-relative paths to configured WPS output directory. This allows the application to easily adjust the exposed result HTTP path according to the service configuration (i.e.: relative to ``weaver.wps_output_dir`` and/or ``weaver.wps_output_url``) and it also avoids rewriting the database job results entry if those settings are changed later on following reboot of the web application. Only references prefixed with ``weaver.wps_output_dir``, ``weaver.wps_output_url`` or a corresponding resolution from ``weaver.wps_output_path`` with ``weaver.url`` will be modified to pseudo-relative paths. Other references (file/URL endpoints that do not correspond to `Weaver`) will be left untouched for literal remote reference. Results that do not correspond to a reference are also unmodified. .. note:: The references are not *real* relative paths (i.e.: starting with ``./``), as those could also be specified as input, and there would be no way to guarantee proper differentiation from paths already handled and stored in the database. Instead, *pseudo-relative* paths employ an explicit *absolute*-like path (i.e.: starting with ``/``) and are assumed to always require to be prefixed by the configured WPS locations (i.e.: ``weaver.wps_output_dir`` or ``weaver.wps_output_url`` based on local or HTTP response context). With this approach, data persistence with mapped volumes into the dockerized `Weaver` service can be placed anywhere at convenience. This is important because sibling docker execution require exact mappings such that volume mount ``/data/path:/data/path`` resolve correctly on both sides (host and image path must be identical). If volumes get remapped differently, ensuring that ``weaver.wps_output_dir`` setting follows the same remapping update will automatically resolve to the proper location for both local references and exposed URL endpoints. :param results: JSON mapping of data results as ``{"<id>": <definition>}`` entries where a reference can be found. :param settings: container to retrieve current application settings. """ wps_url = get_wps_output_url(settings) wps_path = get_wps_output_path(settings) for res in results: ref = res.get("reference") if isinstance(ref, str) and ref: if ref.startswith(wps_url): ref = ref.replace(wps_url, "", 1) if ref.startswith(wps_path): ref = ref.replace(wps_path, "", 1) res["reference"] = ref return results
[docs] def map_locations(job, settings): # type: (Job, SettingsType) -> None """ Maps directory locations between :mod:`pywps` process execution and produced jobs storage. Generates symlink references from the Job UUID to PyWPS UUID results (outputs directory, status and log locations). Update the Job's WPS ID if applicable (job executed locally). Assumes that all results are located under the same reference UUID. """ local_path = get_wps_local_status_location(job.status_location, settings) if not local_path: LOGGER.debug("Not possible to map Job to WPS locations.") return base_dir, status_xml = os.path.split(local_path) job.wps_id = os.path.splitext(status_xml)[0] wps_loc = os.path.join(base_dir, str(job.wps_id)) job_loc = os.path.join(base_dir, str( if wps_loc == job_loc: LOGGER.debug("Job already refers to WPS locations.") return for loc_ext in ["", ".log", ".xml"]: wps_ref = wps_loc + loc_ext job_ref = job_loc + loc_ext if os.path.exists(wps_ref): # possible that there are no results (e.g.: failed job) os.symlink(wps_ref, job_ref)
[docs] def submit_job(request, reference, tags=None): # type: (Request, Union[Service, Process], Optional[List[str]]) -> AnyResponseType """ Generates the job submission from details retrieved in the request. .. seealso:: :func:`submit_job_handler` to provide elements pre-extracted from requests or from other parsing. """ # validate body with expected JSON content and schema if ContentType.APP_JSON not in request.content_type: raise HTTPBadRequest(json={ "code": "InvalidHeaderValue", "name": "Content-Type", "description": f"Request 'Content-Type' header other than '{ContentType.APP_JSON}' not supported.", "value": str(request.content_type) }) try: json_body = request.json_body except Exception as ex: raise HTTPBadRequest(f"Invalid JSON body cannot be decoded for job submission. [{ex}]") # validate context if needed later on by the job for early failure context = get_wps_output_context(request) provider_id = None # None OK if local process_id = None # None OK if remote, but can be found as well if available from WPS-REST path # noqa tags = tags or [] lang = request.accept_language.header_value # can only preemptively check if local process if isinstance(reference, Process): service_url = reference.processEndpointWPS1 process_id = reference.identifier # explicit 'id:version' process revision if available, otherwise simply 'id' visibility = reference.visibility is_workflow = reference.type == ProcessType.WORKFLOW is_local = True tags += "local" support_lang = AcceptLanguage.offers() accepts_lang = request.accept_language # type: AnyAcceptLanguageHeader matched_lang = accepts_lang.lookup(support_lang, default="") or None if lang and not matched_lang: raise HTTPNotAcceptable(f"Requested language [{lang}] not in supported languages [{sorted(support_lang)}].") lang = matched_lang elif isinstance(reference, Service): service_url = reference.url provider_id = process_id = resolve_process_tag(request) visibility = Visibility.PUBLIC is_workflow = False is_local = False tags += "remote" else: # pragma: no cover LOGGER.error("Expected process/service, got: %s", type(reference)) raise TypeError("Invalid process or service reference to execute job.") queries = sd.LaunchJobQuerystring().deserialize(request.params) tags = queries.get("tags", "").split(",") + tags user = request.authenticated_userid # FIXME: consider other methods to provide the user headers = dict(request.headers) settings = get_settings(request) return submit_job_handler(json_body, settings, service_url, provider_id, process_id, is_workflow, is_local, visibility, language=lang, headers=headers, tags=tags, user=user, context=context)
[docs] def submit_job_handler(payload, # type: ProcessExecution settings, # type: SettingsType service_url, # type: str provider=None, # type: Optional[AnyServiceRef] process=None, # type: AnyProcessRef is_workflow=False, # type: bool is_local=True, # type: bool visibility=None, # type: Optional[AnyVisibility] language=None, # type: Optional[str] headers=None, # type: Optional[HeaderCookiesType] tags=None, # type: Optional[List[str]] user=None, # type: Optional[int] context=None, # type: Optional[str] ): # type: (...) -> AnyResponseType """ Submits the job to the Celery worker with provided parameters. Assumes that parameters have been pre-fetched and validated, except for the input payload. """ try: json_body = sd.Execute().deserialize(payload) except colander.Invalid as ex: raise HTTPBadRequest( json=sd.ErrorJsonResponseBodySchema(schema_include=True).deserialize({ "type": "InvalidSchema", "title": "Execute", "detail": "Execution body failed schema validation.", "status": HTTPBadRequest.code, "error": ex.msg, "cause": ex.asdict(), "value": repr_json(ex.value), }) ) db = get_db(settings) # non-local is only a reference, no actual process object to validate provider_id = if isinstance(provider, Service) else provider if process and is_local and not isinstance(process, Process): proc_store = db.get_store(StoreProcesses) process = proc_store.fetch_by_id(process) if process and is_local: validate_process_io(process, json_body) else: LOGGER.warning( "Skipping validation of execution parameters for remote process [%s] on provider [%s]", process, provider_id ) headers = headers or {} if is_local: job_ctl_opts = process.jobControlOptions else: job_ctl_opts = ExecuteControlOption.values() exec_max_wait = settings.get("weaver.execute_sync_max_wait", settings.get("weaver.exec_sync_max_wait")) exec_max_wait = as_int(exec_max_wait, default=20) mode, wait, applied = parse_prefer_header_execute_mode(headers, job_ctl_opts, exec_max_wait) get_header("prefer", headers, pop=True) # don't care about value, just ensure removed with any header container if not applied: # whatever returned is a default, consider 'mode' in body as alternative is_execute_async = ExecuteMode.get(json_body.get("mode")) != ExecuteMode.SYNC # convert auto to async else: # as per # Prefer header not resolved with a valid value should still resume without error is_execute_async = mode != ExecuteMode.SYNC exec_resp = json_body.get("response") subscribers = map_job_subscribers(json_body, settings) store = db.get_store(StoreJobs) # type: StoreJobs job = store.save_job(task_id=Status.ACCEPTED, process=process, service=provider_id, inputs=json_body.get("inputs"), outputs=json_body.get("outputs"), is_local=is_local, is_workflow=is_workflow, access=visibility, user_id=user, context=context, execute_async=is_execute_async, execute_response=exec_resp, custom_tags=tags, accept_language=language, subscribers=subscribers) job.save_log(logger=LOGGER, message="Job task submitted for execution.", status=Status.ACCEPTED, progress=0) job = store.update_job(job) location_url = job.status_url(settings) resp_headers = {"Location": location_url} resp_headers.update(applied) wps_url = clean_ows_url(service_url) result = execute_process.delay(, wps_url=wps_url, headers=headers) # type: CeleryResult LOGGER.debug("Celery pending task [%s] for job [%s].",, if not is_execute_async: LOGGER.debug("Celery task requested as sync if it completes before (wait=%ss)", wait) try: result.wait(timeout=wait) except CeleryTaskTimeoutError: pass if result.ready(): job = store.fetch_by_id( # when sync is successful, it must return the results direct instead of status info # see: if job.status == Status.SUCCEEDED: return get_job_results_response(job, settings, headers=resp_headers) # otherwise return the error status body = job.json(container=settings) body["location"] = location_url resp = get_job_submission_response(body, resp_headers, error=True) return resp else: LOGGER.debug("Celery task requested as sync took too long to complete (wait=%ss). Continue in async.", wait) # sync not respected, therefore must drop it # since both could be provided as alternative preferences, drop only async with limited subset prefer = get_header("Preference-Applied", headers, pop=True) _, _, async_applied = parse_prefer_header_execute_mode({"Prefer": prefer}, [ExecuteMode.ASYNC]) if async_applied: resp_headers.update(async_applied) LOGGER.debug("Celery task submitted to run async.") body = { "jobID":, "processID": job.process, "providerID": provider_id, # dropped by validator if not applicable "status": map_status(Status.ACCEPTED), "location": location_url } resp = get_job_submission_response(body, resp_headers) return resp
[docs] def validate_process_io(process, payload): # type: (Process, ProcessExecution) -> None """ Preemptively verify submitted parameters for execution against expected process definition. Verify missing inputs or obvious type mismatches, but nothing too over-complicated. The ideas behind this function is to avoid unnecessary assignation of :mod:`celery` worker and :term:`Docker` resources that would be guaranteed to fail as soon as the process execution started. This function is **NOT** intended to catch all erroneous inputs, nor validate their values. For example, out-of-range values or unreachable file reference URLs are not guaranteed. However, basic checks such as unacceptable types or cardinality can be performed. Assumes that schema pre-validation was accomplished to minimally guarantee that the structure is valid. :param process: Process description that provides expected inputs and outputs. :param payload: Submitted job execution body. :raises HTTPException: Corresponding error for detected invalid combination of inputs or outputs. """ payload_inputs = convert_input_values_schema(payload.get("inputs", {}), JobInputsOutputsSchema.OLD) payload_outputs = convert_output_params_schema(payload.get("outputs", {}), JobInputsOutputsSchema.OLD) for io_type, io_payload, io_process in [ ("inputs", payload_inputs, process.inputs), ("outputs", payload_outputs, process.outputs), ]: io_payload_set = {get_any_id(io_info) for io_info in io_payload} # can have repeated IDs (list representation) io_process_map = {get_any_id(io_info): io_info for io_info in io_process} # guaranteed unique IDs unknown_ids = set(io_payload_set) - set(io_process_map) if unknown_ids: raise OWSInvalidParameterValue(json={ "code": "InvalidParameterValue", "name": io_type, "description": ( f"Submitted execution {io_type} contain unknown identifiers to the process description. " f"Valid {io_type} identifiers are: {sorted(list(io_process_map))}." ), "value": list(unknown_ids), }) for io_id, io_proc in io_process_map.items(): io_name = f"{io_type}.{io_id}" io_exec = list(filter(lambda _io: get_any_id(_io) == io_id, io_payload)) io_format = io_proc.get("formats", []) # validate format if more strict supported Media-Types are specified # requested format must match with the supported ones by the process # ignore explict any or default plain text representation always available if io_format: io_ctypes = { # field 'type' as Content-Type is only valid in execute payload # during process description, it is used as the data/value type get_field(io_fmt, "mime_type", extra_variations=["type"], default="") for io_fmt in io_exec } io_ctypes = [ctype for ctype in io_ctypes if ctype] io_accept = { get_field(io_fmt, "mime_type", search_variations=True, default="") for io_fmt in io_format } io_accept = [clean_media_type_format(ctype) for ctype in io_accept if ctype] # no format specified explicitly must ensure that the process description has one by default if not io_ctypes: io_default = any(get_field(io_fmt, "default", default=False) for io_fmt in io_format) if not io_default: raise OWSInvalidParameterValue(json={ "code": "InvalidParameterValue", "name": io_name, "description": ( f"Submitted '{io_name}' requires explicit Content-Type specification to" "respect process description that defines no default format." ), "value": { "supportedFormats": list(io_accept), "executionFormats": None, } }) any_types = [ContentType.ANY, ContentType.TEXT_PLAIN] io_accept += any_types if not all(io_fmt in io_accept for io_fmt in io_ctypes): raise OWSInvalidParameterValue(json={ "code": "InvalidParameterValue", "name": io_name, "description": ( f"Submitted '{io_name}' requested Content-Types that do not respect " "supported formats specified by the process description." ), "value": { "supportedFormats": list(io_accept), "executionFormats": list(io_ctypes), } }) if io_type == "inputs": io_min = io_proc["minOccurs"] io_max = io_proc["maxOccurs"] io_len = len(io_exec) if io_len < io_min or (isinstance(io_max, int) and io_len > io_max): raise OWSInvalidParameterValue(json={ "code": "InvalidParameterValue", "name": io_name, "description": f"Submitted '{io_name}' does not respect process description cardinality.", "value": { "minOccurs": io_min, "maxOccurs": io_max, "occurrences": io_len, } })