hallowed-analyst-96384
02/14/2022, 7:52 PMadorable-flower-19656
02/15/2022, 3:08 AMstocky-midnight-78204
02/15/2022, 5:03 AMmysterious-nail-70388
02/15/2022, 6:34 AMmysterious-nail-70388
02/15/2022, 7:31 AMfew-air-56117
02/15/2022, 9:21 AMambitious-guitar-89068
02/15/2022, 9:23 AM` 'Dropping Nifi Processor of type org.apache.nifi.processors.slack.PutSlack, id '
'017718bd-4edc-1e55-534e-ca304519ef4b, name PutSlack from lineage view. This is '
'likely an Ingress or Egress node which may be reading to/writing from external '
'datasets However not currently supported in datahub',`
billions-receptionist-60247
02/15/2022, 10:05 AM{'error': 'Unable to emit metadata to DataHub GMS',
'info': {'exceptionClass': 'com.linkedin.restli.server.RestLiServiceException',
'message': 'java.lang.RuntimeException: Unknown aspect container for entity dataset',
hallowed-gpu-49827
02/15/2022, 10:36 AMstrong-kite-83354
02/15/2022, 2:56 PMnutritious-egg-28432
02/15/2022, 4:07 PMred-napkin-59945
02/15/2022, 5:35 PMLooker View A
and Looker View B
if View B
includes View A
?gentle-optician-51037
02/16/2022, 6:49 AMadorable-flower-19656
02/16/2022, 9:22 AMplain-lion-38626
02/16/2022, 1:31 PMproject1.Dataset2.table2
as an upstream.
but it also has project2.Dataset2.table2
as another upstream.
When using the custom emitter (with UPSERT option) the second project seems to overwrite the first one.
Is this a bug or do I need to query all project and add the lineage upstream afterwards?
lineage_mcp = MetadataChangeProposalWrapper(
entityType="dataset",
changeType=ChangeTypeClass.UPSERT,
entityUrn=builder.make_dataset_urn(platform, fq_table_name, env),
aspectName="upstreamLineage",
aspect=upstream_lineage,)
few-air-56117
02/16/2022, 2:22 PMcuddly-apple-7818
02/16/2022, 3:32 PMfreezing-farmer-89710
02/16/2022, 11:45 PM[2022-02-16 23:04:00,956] ERROR {datahub.entrypoints:125} - File "/usr/local/lib/python3.7/site-packages/datahub/ingestion/api/registry.py", line 84, in _ensure_not_lazy
80 def _ensure_not_lazy(self, key: str) -> Union[Type[T], Exception]:
81 path = self._mapping[key]
82 if isinstance(path, str):
83 try:
--> 84 plugin_class = import_path(path)
85 self.register(key, plugin_class, override=True)
File "/usr/local/lib/python3.7/site-packages/datahub/ingestion/api/registry.py", line 32, in import_path
18 def import_path(path: str) -> Any:
(...)
28 module_name, object_name = path.rsplit(":", 1)
29 else:
30 module_name, object_name = path.rsplit(".", 1)
31
--> 32 item = importlib.import_module(module_name)
33 for attr in object_name.split("."):
File "/usr/local/lib/python3.7/importlib/__init__.py", line 127, in import_module
109 def import_module(name, package=None):
(...)
123 for character in name:
124 if character != '.':
125 break
126 level += 1
--> 127 return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1006, in _gcd_import
File "<frozen importlib._bootstrap>", line 983, in _find_and_load
File "<frozen importlib._bootstrap>", line 967, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 677, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 728, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/usr/local/lib/python3.7/site-packages/datahub/ingestion/source/aws/glue.py", line 27, in <module>
23 from datahub.ingestion.api.source import Source, SourceReport
24 from datahub.ingestion.api.workunit import MetadataWorkUnit
25 from datahub.ingestion.source.aws.aws_common import AwsSourceConfig
26 from datahub.ingestion.source.aws.s3_util import make_s3_urn
--> 27 from datahub.ingestion.source.sql.sql_common import SqlContainerSubTypes
28 from datahub.metadata.com.linkedin.pegasus2avro.common import Status
File "/usr/local/lib/python3.7/site-packages/datahub/ingestion/source/sql/sql_common.py", line 22, in <module>
18 )
19 from urllib.parse import quote_plus
20
21 import pydantic
--> 22 from sqlalchemy import create_engine, inspect
23 from sqlalchemy.engine.reflection import Inspector
---- (full traceback above) ----
File "/usr/local/lib/python3.7/site-packages/datahub/ingestion/api/registry.py", line 84, in _ensure_not_lazy
plugin_class = import_path(path)
File "/usr/local/lib/python3.7/site-packages/datahub/ingestion/api/registry.py", line 32, in import_path
item = importlib.import_module(module_name)
File "/usr/local/lib/python3.7/importlib/__init__.py", line 127, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1006, in _gcd_import
File "<frozen importlib._bootstrap>", line 983, in _find_and_load
File "<frozen importlib._bootstrap>", line 967, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 677, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 728, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/usr/local/lib/python3.7/site-packages/datahub/ingestion/source/aws/glue.py", line 27, in <module>
from datahub.ingestion.source.sql.sql_common import SqlContainerSubTypes
File "/usr/local/lib/python3.7/site-packages/datahub/ingestion/source/sql/sql_common.py", line 22, in <module>
from sqlalchemy import create_engine, inspect
ModuleNotFoundError: No module named 'sqlalchemy'
The above exception was the direct cause of the following exception:
File "/usr/local/lib/python3.7/site-packages/datahub/cli/ingest_cli.py", line 77, in run
67 def run(config: str, dry_run: bool, preview: bool, strict_warnings: bool) -> None:
(...)
73 pipeline_config = load_config_file(config_file)
74
75 try:
76 logger.debug(f"Using config: {pipeline_config}")
--> 77 pipeline = Pipeline.create(pipeline_config, dry_run, preview)
78 except ValidationError as e:
File "/usr/local/lib/python3.7/site-packages/datahub/ingestion/run/pipeline.py", line 175, in create
171 def create(
172 cls, config_dict: dict, dry_run: bool = False, preview_mode: bool = False
173 ) -> "Pipeline":
174 config = PipelineConfig.parse_obj(config_dict)
--> 175 return cls(config, dry_run=dry_run, preview_mode=preview_mode)
File "/usr/local/lib/python3.7/site-packages/datahub/ingestion/run/pipeline.py", line 120, in __init__
105 def __init__(
106 self, config: PipelineConfig, dry_run: bool = False, preview_mode: bool = False
107 ):
(...)
116 preview_mode=preview_mode,
117 )
118
119 source_type = self.config.source.type
--> 120 source_class = source_registry.get(source_type)
121 self.source: Source = source_class.create(
File "/usr/local/lib/python3.7/site-packages/datahub/ingestion/api/registry.py", line 130, in get
115 def get(self, key: str) -> Type[T]:
(...)
126 tp = self._ensure_not_lazy(key)
127 if isinstance(tp, ModuleNotFoundError):
128 raise ConfigurationError(
129 f"{key} is disabled; try running: pip install '{__package_name__}[{key}]'"
--> 130 ) from tp
131 elif isinstance(tp, Exception):
ConfigurationError: glue is disabled; try running: pip install 'acryl-datahub[glue]'
freezing-farmer-89710
02/17/2022, 4:03 AMrich-policeman-92383
02/17/2022, 4:44 AMCaused by: com.linkedin.data.template.TemplateOutputCastException: Invalid URN syntax: Urn entity type should be 'corpuser'.: urn:li:corpUser:datahub
flaky-airplane-82352
02/17/2022, 2:52 PMbillowy-flag-4217
02/17/2022, 8:54 PMinclude:
. This results in an error:
['cannot resolve include: '//path/to.view']
Does anyone know how to resolve this?wide-army-23885
02/17/2022, 10:42 PMstocky-midnight-78204
02/18/2022, 2:32 AMwitty-painting-90923
02/18/2022, 9:33 AM'failures': [{'error': 'Unable to emit metadata to DataHub GMS',
'info': {'exceptionClass': 'com.linkedin.restli.server.RestLiServiceException',
'message': "Parameters of method 'ingest' failed validation with error 'ERROR :: "
'/entity/value/com.linkedin.metadata.snapshot.DatasetSnapshot/aspects/1/com.linkedin.schema.SchemaMetadata/fields/5/type/type/com.linkedin.schema.ArrayType/nestedType '
':: array type is not backed by a DataList\n'
'ERROR :: '
'/entity/value/com.linkedin.metadata.snapshot.DatasetSnapshot/aspects/1/com.linkedin.schema.SchemaMetadata/fields/33/type/type/com.linkedin.schema.ArrayType/nestedType '
':: array type is not backed by a DataList\n'
'ERROR :: '
'/entity/value/com.linkedin.metadata.snapshot.DatasetSnapshot/aspects/1/com.linkedin.schema.SchemaMetadata/fields/37/type/type/com.linkedin.schema.UnionType/nestedTypes '
stale-printer-44316
02/18/2022, 2:36 PMbroad-battery-31188
02/18/2022, 3:11 PMuser
table ?
OperationalError: (pymysql.err.OperationalError) (1142, "SHOW VIEW command denied to user 'datahub'@'<ip address>' for table 'user'")
lemon-hydrogen-83671
02/18/2022, 5:55 PMadd_documentation
transformer out there? I was thinking of adding one that would populate the documentation tab with a template or something for urns gathered in a recipesilly-beach-19296
02/18/2022, 6:02 PMbland-barista-59197
02/19/2022, 1:19 AM