Source code for kfp.dsl._ops_group

# Copyright 2018-2019 The Kubeflow Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#      http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

from typing import Optional, Union
import uuid

from kfp.dsl import _for_loop, _pipeline_param

from . import _container_op
from . import _pipeline


class OpsGroup(object):
    """Represents a logical group of ops and group of OpsGroups.

    This class is the base class for groups of ops, such as ops sharing
    an exit handler, a condition branch, or a loop. This class is not
    supposed to be used by pipeline authors. It is useful for
    implementing a compiler.
    """

    def __init__(self,
                 group_type: str,
                 name: str = None,
                 parallelism: int = None):
        """Create a new instance of OpsGroup.

        Args:
          group_type (str): one of 'pipeline', 'exit_handler', 'condition',
            'for_loop', and 'graph'.
          name (str): name of the opsgroup
          parallelism (int): parallelism for the sub-DAG:s
        """
        #TODO: declare the group_type to be strongly typed
        self.type = group_type
        self.ops = list()
        self.groups = list()
        self.name = name
        self.dependencies = []
        self.parallelism = parallelism
        # recursive_ref points to the opsgroups with the same name if exists.
        self.recursive_ref = None

        self.loop_args = None

    @staticmethod
    def _get_matching_opsgroup_already_in_pipeline(group_type, name):
        """Retrieves the opsgroup when the pipeline already contains it.

        the opsgroup might be already in the pipeline in case of recursive calls.

        Args:
          group_type (str): one of 'pipeline', 'exit_handler', 'condition', and
            'graph'.
          name (str): the name before conversion.
        """
        if not _pipeline.Pipeline.get_default_pipeline():
            raise ValueError('Default pipeline not defined.')
        if name is None:
            return None
        name_pattern = '^' + (group_type + '-' + name + '-').replace(
            '_', '-') + '[\d]+$'
        for ops_group_already_in_pipeline in _pipeline.Pipeline.get_default_pipeline(
        ).groups:
            import re
            if ops_group_already_in_pipeline.type == group_type \
                    and re.match(name_pattern ,ops_group_already_in_pipeline.name):
                return ops_group_already_in_pipeline
        return None

    def _make_name_unique(self):
        """Generate a unique opsgroup name in the pipeline."""
        if not _pipeline.Pipeline.get_default_pipeline():
            raise ValueError('Default pipeline not defined.')

        self.name = (
            self.type + '-' + ('' if self.name is None else self.name + '-') +
            str(_pipeline.Pipeline.get_default_pipeline().get_next_group_id()))
        self.name = self.name.replace('_', '-')

    def __enter__(self):
        if not _pipeline.Pipeline.get_default_pipeline():
            raise ValueError('Default pipeline not defined.')

        self.recursive_ref = self._get_matching_opsgroup_already_in_pipeline(
            self.type, self.name)
        if not self.recursive_ref:
            self._make_name_unique()

        _pipeline.Pipeline.get_default_pipeline().push_ops_group(self)
        return self

    def __exit__(self, *args):
        _pipeline.Pipeline.get_default_pipeline().pop_ops_group()

    def after(self, *ops):
        """Specify explicit dependency on other ops."""
        for op in ops:
            self.dependencies.append(op)
        return self

    def remove_op_recursive(self, op):
        if self.ops and op in self.ops:
            self.ops.remove(op)
        for sub_group in self.groups or []:
            sub_group.remove_op_recursive(op)


[docs]class SubGraph(OpsGroup): TYPE_NAME = 'subgraph' def __init__(self, parallelism: int): if parallelism < 1: raise ValueError( 'SubGraph parallism set to < 1, allowed values are > 0') super(SubGraph, self).__init__(self.TYPE_NAME, parallelism=parallelism)
[docs]class ExitHandler(OpsGroup): """Represents an exit handler that is invoked upon exiting a group of ops. Args: exit_op: An operator invoked at exiting a group of ops. Raises: ValueError: Raised if the exit_op is invalid. Example: :: exit_op = ContainerOp(...) with ExitHandler(exit_op): op1 = ContainerOp(...) op2 = ContainerOp(...) """ def __init__( self, exit_op: _container_op.ContainerOp, name: Optional[str] = None, ): super(ExitHandler, self).__init__('exit_handler') # Use user specified name as display name directly, instead of passing # it to the super class constructor. self.display_name = name if exit_op.dependent_names: raise ValueError('exit_op cannot depend on any other ops.') # Removing exit_op form any group _pipeline.Pipeline.get_default_pipeline().remove_op_from_groups(exit_op) # Setting is_exit_handler since the compiler might be using this attribute. # TODO: Check that it's needed exit_op.is_exit_handler = True self.exit_op = exit_op
[docs]class Condition(OpsGroup): """Represents an condition group with a condition. Args: condition (ConditionOperator): the condition. name (str): name of the condition Example: :: with Condition(param1=='pizza', '[param1 is pizza]'): op1 = ContainerOp(...) op2 = ContainerOp(...) """ def __init__(self, condition, name=None): super(Condition, self).__init__('condition', name) self.condition = condition
class Graph(OpsGroup): """Graph DAG with inputs, recursive_inputs, and outputs. This is not used directly by the users but auto generated when the graph_component decoration exists Args: name: Name of the graph. """ def __init__(self, name): super(Graph, self).__init__(group_type='graph', name=name) self.inputs = [] self.outputs = {} self.dependencies = []
[docs]class ParallelFor(OpsGroup): """Represents a parallel for loop over a static set of items. Example: In this case :code:`op1` would be executed twice, once with case :code:`args=['echo 1']` and once with case :code:`args=['echo 2']`:: with dsl.ParallelFor([{'a': 1, 'b': 10}, {'a': 2, 'b': 20}]) as item: op1 = ContainerOp(..., args=['echo {}'.format(item.a)]) op2 = ContainerOp(..., args=['echo {}'.format(item.b]) """ TYPE_NAME = 'for_loop' def __init__(self, loop_args: Union[_for_loop.ItemList, _pipeline_param.PipelineParam], parallelism: int = None): if parallelism and parallelism < 1: raise ValueError( 'ParallelFor parallism set to < 1, allowed values are > 0') self.items_is_pipeline_param = isinstance(loop_args, _pipeline_param.PipelineParam) super().__init__(self.TYPE_NAME, parallelism=parallelism) if self.items_is_pipeline_param: loop_args = _for_loop.LoopArguments.from_pipeline_param(loop_args) elif not self.items_is_pipeline_param and not isinstance( loop_args, _for_loop.LoopArguments): # we were passed a raw list, wrap it in loop args loop_args = _for_loop.LoopArguments( loop_args, code=str(_pipeline.Pipeline.get_default_pipeline() .get_next_group_id()), ) self.loop_args = loop_args def __enter__(self) -> _for_loop.LoopArguments: _ = super().__enter__() return self.loop_args