Skip to content
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 0 additions & 2 deletions cylc/flow/data_messages.proto
Original file line number Diff line number Diff line change
Expand Up @@ -170,7 +170,6 @@ message PbTask {
optional PbMeta meta = 4;
optional float mean_elapsed_time = 5;
optional int32 depth = 6;
repeated string proxies = 7;
repeated string namespace = 8;
repeated string parents = 9;
optional string first_parent = 10;
Expand Down Expand Up @@ -251,7 +250,6 @@ message PbFamily {
optional string name = 3;
optional PbMeta meta = 4;
optional int32 depth = 5;
repeated string proxies = 6;
repeated string parents = 7;
repeated string child_tasks = 8;
repeated string child_families = 9;
Expand Down
94 changes: 47 additions & 47 deletions cylc/flow/data_messages_pb2.py

Large diffs are not rendered by default.

12 changes: 4 additions & 8 deletions cylc/flow/data_messages_pb2.pyi
Original file line number Diff line number Diff line change
Expand Up @@ -228,14 +228,13 @@ class PbJob(_message.Message):
def __init__(self, stamp: _Optional[str] = ..., id: _Optional[str] = ..., submit_num: _Optional[int] = ..., state: _Optional[str] = ..., task_proxy: _Optional[str] = ..., submitted_time: _Optional[str] = ..., started_time: _Optional[str] = ..., finished_time: _Optional[str] = ..., job_id: _Optional[str] = ..., job_runner_name: _Optional[str] = ..., execution_time_limit: _Optional[float] = ..., platform: _Optional[str] = ..., job_log_dir: _Optional[str] = ..., name: _Optional[str] = ..., cycle_point: _Optional[str] = ..., messages: _Optional[_Iterable[str]] = ..., runtime: _Optional[_Union[PbRuntime, _Mapping]] = ..., estimated_finish_time: _Optional[str] = ...) -> None: ...

class PbTask(_message.Message):
__slots__ = ("stamp", "id", "name", "meta", "mean_elapsed_time", "depth", "proxies", "namespace", "parents", "first_parent", "runtime")
__slots__ = ("stamp", "id", "name", "meta", "mean_elapsed_time", "depth", "namespace", "parents", "first_parent", "runtime")
STAMP_FIELD_NUMBER: _ClassVar[int]
ID_FIELD_NUMBER: _ClassVar[int]
NAME_FIELD_NUMBER: _ClassVar[int]
META_FIELD_NUMBER: _ClassVar[int]
MEAN_ELAPSED_TIME_FIELD_NUMBER: _ClassVar[int]
DEPTH_FIELD_NUMBER: _ClassVar[int]
PROXIES_FIELD_NUMBER: _ClassVar[int]
NAMESPACE_FIELD_NUMBER: _ClassVar[int]
PARENTS_FIELD_NUMBER: _ClassVar[int]
FIRST_PARENT_FIELD_NUMBER: _ClassVar[int]
Expand All @@ -246,12 +245,11 @@ class PbTask(_message.Message):
meta: PbMeta
mean_elapsed_time: float
depth: int
proxies: _containers.RepeatedScalarFieldContainer[str]
namespace: _containers.RepeatedScalarFieldContainer[str]
parents: _containers.RepeatedScalarFieldContainer[str]
first_parent: str
runtime: PbRuntime
def __init__(self, stamp: _Optional[str] = ..., id: _Optional[str] = ..., name: _Optional[str] = ..., meta: _Optional[_Union[PbMeta, _Mapping]] = ..., mean_elapsed_time: _Optional[float] = ..., depth: _Optional[int] = ..., proxies: _Optional[_Iterable[str]] = ..., namespace: _Optional[_Iterable[str]] = ..., parents: _Optional[_Iterable[str]] = ..., first_parent: _Optional[str] = ..., runtime: _Optional[_Union[PbRuntime, _Mapping]] = ...) -> None: ...
def __init__(self, stamp: _Optional[str] = ..., id: _Optional[str] = ..., name: _Optional[str] = ..., meta: _Optional[_Union[PbMeta, _Mapping]] = ..., mean_elapsed_time: _Optional[float] = ..., depth: _Optional[int] = ..., namespace: _Optional[_Iterable[str]] = ..., parents: _Optional[_Iterable[str]] = ..., first_parent: _Optional[str] = ..., runtime: _Optional[_Union[PbRuntime, _Mapping]] = ...) -> None: ...

class PbPollTask(_message.Message):
__slots__ = ("local_proxy", "workflow", "remote_proxy", "req_state", "graph_string")
Expand Down Expand Up @@ -399,13 +397,12 @@ class PbTaskProxy(_message.Message):
def __init__(self, stamp: _Optional[str] = ..., id: _Optional[str] = ..., task: _Optional[str] = ..., state: _Optional[str] = ..., cycle_point: _Optional[str] = ..., depth: _Optional[int] = ..., job_submits: _Optional[int] = ..., outputs: _Optional[_Mapping[str, PbOutput]] = ..., namespace: _Optional[_Iterable[str]] = ..., prerequisites: _Optional[_Iterable[_Union[PbPrerequisite, _Mapping]]] = ..., jobs: _Optional[_Iterable[str]] = ..., first_parent: _Optional[str] = ..., name: _Optional[str] = ..., is_held: bool = ..., edges: _Optional[_Iterable[str]] = ..., ancestors: _Optional[_Iterable[str]] = ..., flow_nums: _Optional[str] = ..., external_triggers: _Optional[_Mapping[str, PbTrigger]] = ..., xtriggers: _Optional[_Mapping[str, PbTrigger]] = ..., is_queued: bool = ..., is_runahead: bool = ..., flow_wait: bool = ..., runtime: _Optional[_Union[PbRuntime, _Mapping]] = ..., graph_depth: _Optional[int] = ..., is_retry: bool = ..., is_wallclock: bool = ..., is_xtriggered: bool = ...) -> None: ...

class PbFamily(_message.Message):
__slots__ = ("stamp", "id", "name", "meta", "depth", "proxies", "parents", "child_tasks", "child_families", "first_parent", "runtime")
__slots__ = ("stamp", "id", "name", "meta", "depth", "parents", "child_tasks", "child_families", "first_parent", "runtime")
STAMP_FIELD_NUMBER: _ClassVar[int]
ID_FIELD_NUMBER: _ClassVar[int]
NAME_FIELD_NUMBER: _ClassVar[int]
META_FIELD_NUMBER: _ClassVar[int]
DEPTH_FIELD_NUMBER: _ClassVar[int]
PROXIES_FIELD_NUMBER: _ClassVar[int]
PARENTS_FIELD_NUMBER: _ClassVar[int]
CHILD_TASKS_FIELD_NUMBER: _ClassVar[int]
CHILD_FAMILIES_FIELD_NUMBER: _ClassVar[int]
Expand All @@ -416,13 +413,12 @@ class PbFamily(_message.Message):
name: str
meta: PbMeta
depth: int
proxies: _containers.RepeatedScalarFieldContainer[str]
parents: _containers.RepeatedScalarFieldContainer[str]
child_tasks: _containers.RepeatedScalarFieldContainer[str]
child_families: _containers.RepeatedScalarFieldContainer[str]
first_parent: str
runtime: PbRuntime
def __init__(self, stamp: _Optional[str] = ..., id: _Optional[str] = ..., name: _Optional[str] = ..., meta: _Optional[_Union[PbMeta, _Mapping]] = ..., depth: _Optional[int] = ..., proxies: _Optional[_Iterable[str]] = ..., parents: _Optional[_Iterable[str]] = ..., child_tasks: _Optional[_Iterable[str]] = ..., child_families: _Optional[_Iterable[str]] = ..., first_parent: _Optional[str] = ..., runtime: _Optional[_Union[PbRuntime, _Mapping]] = ...) -> None: ...
def __init__(self, stamp: _Optional[str] = ..., id: _Optional[str] = ..., name: _Optional[str] = ..., meta: _Optional[_Union[PbMeta, _Mapping]] = ..., depth: _Optional[int] = ..., parents: _Optional[_Iterable[str]] = ..., child_tasks: _Optional[_Iterable[str]] = ..., child_families: _Optional[_Iterable[str]] = ..., first_parent: _Optional[str] = ..., runtime: _Optional[_Union[PbRuntime, _Mapping]] = ...) -> None: ...

class PbFamilyProxy(_message.Message):
__slots__ = ("stamp", "id", "cycle_point", "name", "family", "state", "depth", "first_parent", "child_tasks", "child_families", "is_held", "ancestors", "states", "state_totals", "is_held_total", "is_queued", "is_queued_total", "is_runahead", "is_runahead_total", "runtime", "graph_depth", "is_retry", "is_wallclock", "is_xtriggered")
Expand Down
19 changes: 0 additions & 19 deletions cylc/flow/data_store_mgr.py
Original file line number Diff line number Diff line change
Expand Up @@ -494,9 +494,6 @@ def apply_delta(key, delta, data):
# The suppression of key/value errors is to avoid
# elements and their relationships missing on reload.
if key == TASK_PROXIES:
# remove relationship from task
with suppress(KeyError, ValueError):
data[TASKS][data[key][del_id].task].proxies.remove(del_id)
# remove relationship from parent/family
with suppress(KeyError, ValueError):
data[FAMILY_PROXIES][
Expand All @@ -506,10 +503,6 @@ def apply_delta(key, delta, data):
with suppress(KeyError, ValueError):
getattr(data[WORKFLOW], key).remove(del_id)
elif key == FAMILY_PROXIES:
with suppress(KeyError, ValueError):
data[FAMILIES][
data[key][del_id].family
].proxies.remove(del_id)
with suppress(KeyError, ValueError):
data[FAMILY_PROXIES][
data[key][del_id].first_parent
Expand Down Expand Up @@ -1407,13 +1400,6 @@ def generate_ghost_task(

self.added[TASK_PROXIES][tp_id] = tproxy
getattr(self.updated[WORKFLOW], TASK_PROXIES).append(tp_id)
self.updated[TASKS].setdefault(
t_id,
PbTask(
stamp=f'{t_id}@{update_time}',
id=t_id,
)
).proxies.append(tp_id)
self.generate_ghost_family(tproxy.first_parent, child_task=tp_id)
self.state_update_families.add(tproxy.first_parent)

Expand Down Expand Up @@ -1538,11 +1524,6 @@ def generate_ghost_family(self, fp_id, child_fam=None, child_task=None):

self.added[FAMILY_PROXIES][fp_id] = fp_delta
fp_parent = fp_delta
# Add ref ID to family element
f_delta = PbFamily(id=fam.id, stamp=f'{fam.id}@{update_time}')
f_delta.proxies.append(fp_id)
self.updated[FAMILIES].setdefault(
fam.id, PbFamily(id=fam.id)).MergeFrom(f_delta)
# Add ref ID to workflow element
getattr(self.updated[WORKFLOW], FAMILY_PROXIES).append(fp_id)
# Generate this families parent if it not root.
Expand Down
16 changes: 0 additions & 16 deletions cylc/flow/network/schema.py
Original file line number Diff line number Diff line change
Expand Up @@ -1011,14 +1011,6 @@ class Meta:
depth = Int(
description='The family inheritance depth.'
)
proxies = graphene.List(
lambda: TaskProxy,
description="Associated cycle point proxies",
args=PROXY_ARGS,
strip_null=STRIP_NULL_DEFAULT,
delta_store=DELTA_STORE_DEFAULT,
delta_type=DELTA_TYPE_DEFAULT,
resolver=get_nodes_by_ids)
parents = graphene.List(
lambda: Family,
description="Family definition parent.",
Expand Down Expand Up @@ -1261,14 +1253,6 @@ class Meta:
meta = Field(NodeMeta)
runtime = Field(Runtime)
depth = Int()
proxies = graphene.List(
lambda: FamilyProxy,
description='Associated family proxy instances.',
args=PROXY_ARGS,
strip_null=STRIP_NULL_DEFAULT,
delta_store=DELTA_STORE_DEFAULT,
delta_type=DELTA_TYPE_DEFAULT,
resolver=get_nodes_by_ids)
parents = graphene.List(
lambda: Family,
description='Families that this family directly inherits from.',
Expand Down
Loading