Skip to content

Commit 2f78ab1

Browse files
authored
fix(propagation/doc): Don't fail in doc propagation if origin is missing (#161)
1 parent 649f1d4 commit 2f78ab1

File tree

4 files changed

+16
-40
lines changed

4 files changed

+16
-40
lines changed

.github/workflows/build-and-test.yml

+2-2
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ jobs:
3131
- name: Gradle build (and test)
3232
run: |
3333
./gradlew build
34-
- uses: actions/upload-artifact@v3
34+
- uses: actions/upload-artifact@v4
3535
if: always()
3636
with:
3737
name: Test Results (build)
@@ -44,7 +44,7 @@ jobs:
4444
runs-on: ubuntu-latest
4545
steps:
4646
- name: Upload
47-
uses: actions/upload-artifact@v3
47+
uses: actions/upload-artifact@v4
4848
with:
4949
name: Event File
5050
path: ${{ github.event_path }}

.github/workflows/datahub-actions-docker.yml

+4-3
Original file line numberDiff line numberDiff line change
@@ -152,7 +152,7 @@ jobs:
152152
run: docker image save -o image.tar ${{ steps.docker_meta_slim.outputs.tags }}
153153
- name: Upload artifact
154154
if: needs.setup.outputs.publish != 'true'
155-
uses: actions/upload-artifact@v3
155+
uses: actions/upload-artifact@v4
156156
with:
157157
name: docker-image
158158
path: image.tar
@@ -201,7 +201,7 @@ jobs:
201201
uses: actions/checkout@v4
202202
- name: Download artifact (if not publishing)
203203
if: needs.setup.outputs.publish != 'true'
204-
uses: actions/download-artifact@v3
204+
uses: actions/download-artifact@v4
205205
with:
206206
name: docker-image
207207
- name: Load Docker image (if not publishing)
@@ -214,6 +214,7 @@ jobs:
214214
uses: aquasecurity/trivy-action@master
215215
env:
216216
TRIVY_OFFLINE_SCAN: true
217+
TRIVY_DB_REPOSITORY: public.ecr.aws/aquasecurity/trivy-db:2,ghcr.io/aquasecurity/trivy-db:2
217218
with:
218219
image-ref: acryldata/datahub-actions-slim:${{ needs.setup.outputs.unique_tag }}
219220
format: "template"
@@ -252,7 +253,7 @@ jobs:
252253
cache: "pip"
253254
- name: Download artifact (if not publishing)
254255
if: needs.setup.outputs.publish != 'true'
255-
uses: actions/download-artifact@v3
256+
uses: actions/download-artifact@v4
256257
with:
257258
name: docker-image
258259
- name: Load Docker image (if not publishing)

datahub-actions/src/datahub_actions/plugin/action/propagation/docs/propagation_action.py

+5-1
Original file line numberDiff line numberDiff line change
@@ -305,7 +305,11 @@ def process_schema_field_documentation(
305305
propagation_relationships = self.get_propagation_relationships(
306306
entity_type="schemaField", source_details=source_details_parsed
307307
)
308-
origin_entity = source_details_parsed.origin
308+
origin_entity = (
309+
source_details_parsed.origin
310+
if source_details_parsed.origin
311+
else entity_urn
312+
)
309313
if old_docs is None or not old_docs.documentations:
310314
return DocPropagationDirective(
311315
propagate=True,

smoke-test/tests/actions/doc_propagation/test_propagation.py

+5-34
Original file line numberDiff line numberDiff line change
@@ -5,11 +5,11 @@
55
import uuid
66
from contextlib import contextmanager
77
from pathlib import Path
8-
from typing import Any, Dict, Iterable, List, Tuple
8+
from typing import Any, Dict, Iterator, List, Tuple
99

1010
import datahub.metadata.schema_classes as models
11-
from pydantic import BaseModel
1211
import pytest
12+
import tenacity
1313
from datahub.api.entities.dataset.dataset import Dataset
1414
from datahub.emitter.mce_builder import make_schema_field_urn
1515
from datahub.emitter.mcp import MetadataChangeProposalWrapper
@@ -19,11 +19,7 @@
1919
from datahub.ingestion.sink.file import FileSink, FileSinkConfig
2020
from datahub.utilities.urns.urn import Urn
2121
from jinja2 import Template
22-
import tenacity
23-
from datahub_actions.plugin.action.propagation.docs.propagation_action import (
24-
DocPropagationConfig,
25-
)
26-
22+
from pydantic import BaseModel
2723

2824
from tests.utils import (
2925
delete_urns_from_file,
@@ -111,7 +107,7 @@ def action_env_vars(pytestconfig) -> ActionTestEnv:
111107
key, value = line.split("=", 1)
112108
env_vars[key] = value
113109

114-
return ActionTestEnv(**env_vars)
110+
return ActionTestEnv.parse_obj(env_vars)
115111

116112

117113
@pytest.fixture(scope="function")
@@ -164,31 +160,6 @@ def test_resources_dir(root_dir):
164160
return Path(root_dir) / "tests" / "actions" / "doc_propagation" / "resources"
165161

166162

167-
@pytest.fixture(scope="function")
168-
def ingest_cleanup_data_function(request, test_resources_dir, graph, test_id):
169-
@contextmanager
170-
def _ingest_cleanup_data(template_file="datasets_template.yaml"):
171-
new_file, filename = tempfile.mkstemp(suffix=f"_{test_id}.json")
172-
try:
173-
template_path = Path(test_resources_dir) / template_file
174-
all_urns = create_test_data(filename, template_path, test_id)
175-
print(
176-
f"Ingesting datasets test data for test_id: {test_id} using template: {template_file}"
177-
)
178-
ingest_file_via_rest(filename)
179-
yield all_urns
180-
finally:
181-
if DELETE_AFTER_TEST:
182-
print(f"Removing test data for test_id: {test_id}")
183-
delete_urns_from_file(filename)
184-
for urn in all_urns:
185-
graph.delete_entity(urn, hard=True)
186-
wait_for_writes_to_sync()
187-
os.remove(filename)
188-
189-
return _ingest_cleanup_data
190-
191-
192163
@pytest.fixture(scope="function")
193164
def ingest_cleanup_data(ingest_cleanup_data_function):
194165
"""
@@ -261,7 +232,7 @@ def large_fanout_graph_function(graph: DataHubGraph):
261232
@contextmanager
262233
def _large_fanout_graph(
263234
test_id: str, max_fanout: int
264-
) -> Iterable[Tuple[str, List[str]]]:
235+
) -> Iterator[Tuple[str, List[str]]]:
265236
max_index = max_fanout + 1
266237
all_urns = []
267238
dataset_base_name = f"large_fanout_dataset_{test_id}"

0 commit comments

Comments
 (0)