white-horse-97256
02/13/2023, 7:14 PMThe datum UpstreamLineageClass({'upstreams': [UpstreamClass({'auditStamp': AuditStampClass({'time': 0, 'actor': 'urn:li:corpuser:unknown', 'impersonator': None, 'message': None}), 'created': None, 'dataset': 'urn:li:dataset:(urn:li:dataPlatform:neo4j,labels.Asset,STG)', 'type': 'TRANSFORMED', 'properties': None})], 'fineGrainedLineages': [FineGrainedLineageClass({'upstreamType': 'FIELD_SET', 'upstreams': ['urn:li:schemaField:(urn:li:dataset:(urn:li:dataPlatform:neo4j,labels.Asset,STG),account_id)'], 'downstreamType': 'NONE', 'downstreams': [], 'transformOperation': None, 'confidenceScore': 1.0})]}) is not an example of the schema.
powerful-telephone-2424
02/13/2023, 9:32 PMcold-airport-17919
02/13/2023, 9:57 PMbland-lighter-26751
02/14/2023, 12:03 AMambitious-notebook-45027
02/14/2023, 2:11 AMFAILED: SemanticException [Error 10056]:
Queries against partitioned tables without a partition filter are disabled for safety reasons.
If you know what you are doing, please set hive.strict.checks.no.partition.
filter to false and make sure that hive.mapred.mode is not set to 'strict' to proceed.
Note that you may get errors or incorrect results if you make a mistake while using some of the unsafe features.
No partition predicate for Alias "lubian" Table "lubian"
how can i do?@Mayuri Nhallowed-shampoo-52722
02/14/2023, 5:10 AMplain-cricket-83456
02/14/2023, 7:26 AMshy-hairdresser-85182
02/14/2023, 9:18 AMcolossal-smartphone-90274
02/14/2023, 12:24 PMdef get_all_reports(self) -> List[Any]:
"""
Fetch all Reports from PowerBI Report Server
"""
report_types_mapping: Dict[str, Any] = {
Constant.REPORTS: Report,
Constant.MOBILE_REPORTS: MobileReport,
Constant.LINKED_REPORTS: LinkedReport,
Constant.POWERBI_REPORTS: PowerBiReport,
}
On the PoC version of DataHub, I removed the MOBILE_REPORTS line of the code snippet and the ingest worked again however I will need a different strategy for my OpenShift deployment. Has anyone else had this issue with the ingest?rich-pager-68736
02/14/2023, 2:43 PMdazzling-microphone-98929
02/14/2023, 2:47 PMlemon-scooter-69730
02/14/2023, 5:22 PM('Failed to load service account credentials from /tmp/tmpuvp2cqms', ValueError('Could not deserialize key data. The data may be in an incorrect format, it may be encrypted with an unsupported algorithm, or it may be an unsupported key type (e.g. EC curves with explicit parameters).', [_OpenSSLErrorWithText(code=503841036, lib=60, reason=524556, reason_text=b'error:1E08010C:DECODER routines::unsupported')]))
tall-caravan-42586
02/14/2023, 5:53 PMtall-caravan-42586
02/14/2023, 5:54 PMfancy-crayon-39356
02/14/2023, 6:57 PMdatahub cli
on version v0.10.0
.
Digging into this problem I've found this PR: https://github.com/datahub-project/datahub/pull/7063/files that changed the DBTColumn
name from catalog_column["name"].lower()
to catalog_column["name"]
. Essentially, making the column URN the same as in the catalog (which comes from snowflake and, in that case, its uppercased). The problem is that in the Snowflake recipe we are lowercasing urns by default (convert_urns_to_lowercase=True
), causing the mismatch.
What is the standard going forward here? Are we sticking to lowercase urns to ensure cross-platform compatibility or DBT will use whatever is defined in the catalog? I'm happy to submit a PR to maybe introduce a convert_urns_to_lowercase
flag to the DBT recipe as well, if that's the standard going forward.bland-barista-59197
02/14/2023, 7:00 PMenough-lamp-79907
02/14/2023, 7:21 PMsource:
type: s3
config:
path_specs:
- include: <s3://test/dumps/kafka/order/daily/{partition_key[0]}={partition[0]}/*.parquet>
aws_config:
aws_profile: dev
aws_region: eu-central-1
env: "dev"
profiling:
enabled: false
white-horse-97256
02/14/2023, 9:39 PMpolite-actor-701
02/15/2023, 1:55 AMcalm-jewelry-98911
02/15/2023, 3:21 AMCorpUserStatus
as suspended using the MCP below (and subsequently emitting it) ->
In [47]: import time
...: mcp2 = MetadataChangeProposalWrapper(
...: entityType="corpuser",
...: changeType=ChangeTypeClass.UPSERT,
...: entityUrn=make_user_urn('apanwar'),
...: aspectName=CorpUserStatusClass.get_aspect_name(),
...: aspect=CorpUserStatusClass(
...: status='SUSPENDED',
...: lastModified=AuditStampClass(
...: time=int(time.time()*1000),
...: actor='urn:li:corpuser:datahub'
...: )
...: ),
...: )
The GMS logs have captured an error related to this -
Caused by: java.lang.IllegalArgumentException: No enum constant com.linkedin.datahub.graphql.generated.CorpUserStatus.SUSPENDED
at java.base/java.lang.Enum.valueOf(Enum.java:240)
at com.linkedin.datahub.graphql.generated.CorpUserStatus.valueOf(CorpUserStatus.java:6)
at com.linkedin.datahub.graphql.types.corpuser.mappers.CorpUserStatusMapper.apply(CorpUserStatusMapper.java:19)
at com.linkedin.datahub.graphql.types.corpuser.mappers.CorpUserStatusMapper.map(CorpUserStatusMapper.java:13)
at com.linkedin.datahub.graphql.types.corpuser.mappers.CorpUserMapper.lambda$apply$3(CorpUserMapper.java:64)
at com.linkedin.datahub.graphql.types.common.mappers.util.MappingHelper.mapToResult(MappingHelper.java:22)
at com.linkedin.datahub.graphql.types.corpuser.mappers.CorpUserMapper.apply(CorpUserMapper.java:63)
at com.linkedin.datahub.graphql.types.corpuser.mappers.CorpUserMapper.map(CorpUserMapper.java:46)
at com.linkedin.datahub.graphql.types.corpuser.CorpUserType.lambda$batchLoad$0(CorpUserType.java:95)
at java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:195)
at java.base/java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1655)
at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
at java.base/java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:913)
at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.base/java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:578)
at com.linkedin.datahub.graphql.types.corpuser.CorpUserType.batchLoad(CorpUserType.java:96)
... 18 common frames omitted
Was wondering if I missed something here? I thought SUSPENDED would be a valid value for the CorpUserStatus.status
as mentioned in the schema class' getter and setter ->plain-nest-12882
02/15/2023, 5:29 AMnumerous-account-62719
02/15/2023, 8:46 AMrich-policeman-92383
02/15/2023, 10:05 AMbroad-wire-76841
02/15/2023, 10:49 AMripe-eye-60209
02/15/2023, 6:31 PMlimited-forest-73733
02/15/2023, 6:50 PMwhite-horse-97256
02/15/2023, 7:34 PMcalm-jewelry-98911
02/15/2023, 7:55 PMCorpUserStatus
but faced issues as described in the OG question) -
https://datahubspace.slack.com/archives/CUMUWQU66/p1676431303025729silly-dog-87292
02/15/2023, 7:57 PMwhite-horse-97256
02/15/2023, 10:28 PM