ripe-apple-36185
10/11/2022, 4:21 PMconvert_urns_to_lowercase: false in the recipe).
great expectations is converting the URN components to lower case. Is there a way to have DataHubValidationAction set the URNs to uppercase?hundreds-photographer-13496
10/12/2022, 5:07 AMconvert_urns_to_lowercase: True ? Do you use databases/schemas/tables with exactly same literals but different case so that their urns might collide if converted to lowercase ?ripe-apple-36185
10/12/2022, 12:43 PMconvert_urns_to_lowercase: false to preserve how it is in Snowflake and be able to mesh metadata provided from different sources.
Do you know if there is a way to emmit the output from the Do you know if there is a way to emmit the output from DataHubValidationAction to a file? It will help me understand what ios being sent.
The URNs I have in DH are: urn:li:dataset:(urn:li:dataPlatform:snowflake,RAW.ANALYTICS.STG_CUSTOMERS,PROD)
When I change to lower case and have URNs like urn:li:dataset:(urn:li:dataPlatform:snowflake,raw.analytics.stg_customers,PROD), I see the GE results in.
I tried changing requires_name_normalize = True in snowdialect.py, but that seems to only change the field names.ripe-apple-36185
10/12/2022, 12:45 PMhundreds-photographer-13496
10/13/2022, 6:13 AMDo you know if there is a way to emmit the output fromIts not possible at the moment. However if you are using python script to run the checkpoint, I can help with how to enable debug logs containing emitted output. From your observations, it looks liketo a file?DataHubValidationAction
urn:li:dataset:(urn:li:dataPlatform:snowflake,raw.analytics.stg_customers,PROD) is the urn constructed by DatahubValidationAction.
I am curious what was the urn was when you changed requires_name_normalize = False in snowdialect.py , as in DataHubValidationAction , when generating urn, only database name is explicitly converted to lowecase whereas schema and table name are read from GE batch spec, as is.ripe-apple-36185
10/13/2022, 11:47 AMhundreds-photographer-13496
10/13/2022, 12:35 PMimport logging
import great_expectations as ge
datahub_logger = logging.getLogger("datahub")
datahub_logger.setLevel(logging.DEBUG)
datahub_logger.addHandler(logging.StreamHandler())
context = ge.get_context()
context.run_checkpoint(checkpoint_name="<name of checkpoint>")
I have created this PR to be able to display debug logs just by setting environment variable DATAHUB_DEBUG=True. so in future, we should be able to do this without writing code 🙂ripe-apple-36185
10/13/2022, 12:38 PMripe-apple-36185
10/13/2022, 1:48 PMurn:li:dataset:(urn:li:dataPlatform:snowflake,raw.analytics.stg_customers,PROD)hundreds-photographer-13496
10/14/2022, 7:59 AMhundreds-photographer-13496
10/14/2022, 8:16 AM