magnificent-plumber-63682
06/16/2023, 9:59 AMgreat-notebook-53658
06/19/2023, 2:26 AMshy-dog-84302
06/19/2023, 12:17 PM[2023-06-19 12:07:32,796] INFO {datahub.ingestion.source.bigquery_v2.bigquery:474} - Processing project: xyz-project
[2023-06-19 12:07:33,008] WARNING {datahub.ingestion.source.bigquery_v2.bigquery:589} - No dataset found in xyz-project. Either there are no datasets in this project or missing bigquery.datasets.get permission. You can assign predefined roles/bigquery.metadataViewer role to your service account.
[2023-06-19 12:07:33,008] INFO {datahub.ingestion.source.bigquery_v2.bigquery_report:95} - Time spent in stage <xyz-project: Metadata Extraction at 2023-06-19 12:07:32.796768+00:00>: 0.21 seconds
millions-city-84223
06/19/2023, 12:40 PMbland-application-65186
06/19/2023, 2:30 PMstrong-diamond-4751
06/19/2023, 7:59 PMicy-zoo-92866
06/20/2023, 7:18 AMsource:
type: superset
config:
connect_uri: '<https://superset-xx.xx.xx/>'
username: xxx
password: xxx
provider: db
env: xxx
Request/Auth is successful but we are not getting any dashboards or charts back. When I login with same user and pwd to superset I can see all the charts
what can be the issue..?
TIAminiature-hair-20451
06/20/2023, 8:00 AMjolly-airline-17196
06/20/2023, 11:52 AMraise Exception(f"Failed to process {path}")
Exception: Failed to process /home/datahub/students.json
ancient-queen-15575
06/20/2023, 1:21 PMlively-raincoat-33818
06/20/2023, 3:48 PMlimited-cricket-18852
06/20/2023, 4:54 PMDatasets/ prod/ databricks/ my_workspace/ global-euwest/ my_catalog/ some_layer/ my_beautiful_table
, however the my_workspace
and global-euwest
are not interesting for me. Is there a way to ingest without these information?
Thanks!bumpy-hamburger-47757
06/20/2023, 7:52 PMDataHubGraph.get_urns_by_filter()
and it's returning partial name matches for dataset name and column names. For example, if my query is "test_table"
, it will return any datasets with the words test
or table
in the dataset name or columns (for example, a dataset named users_table
or a column named test_value
will match). Thanks!average-nail-72662
06/20/2023, 9:15 PMbland-orange-13353
06/21/2023, 12:47 AMeager-monitor-4683
06/21/2023, 3:21 AMrefined-gold-30439
06/21/2023, 8:18 AMsource:
type: looker
config:
base_url: '<https://lookerstudio.google.com/>'
client_id: '${looker_client_id}'
stateful_ingestion:
enabled: true
client_secret: '${looker_client_secret}'
• Error
[2023-06-21 08:17:34,423] INFO {looker_sdk.rtl.requests_transport:72} - POST(<https://lookerstudio.google.com//api/4.0/login>)
[2023-06-21 08:17:35,243] ERROR {datahub.entrypoints:199} - Command failed: Failed to configure the source (looker): Failed to connect/authenticate with looker - check your configuration: )]}'
{"errorStatus":{"code":9}}
gifted-bird-57147
06/21/2023, 9:10 AMsource:
type: athena
config:
# Coordinates
aws_region: eu-west-1
work_group: ${ATHENA_WG_PROD_BDV}
username: ${ATHENA_USER_BDV}
password: ${ATHENA_PW_BDV}
query_result_location: ${ATHENA_QL_BDV}
## vanwege een bug in de athena ingestion moeten we de database opgeven.
## Daarom aparte scripts per database (want je kunt maar 1 database per script specificeren...)
database: "bdv-prod-topdesk-transformed"
# Options
#s3_staging_dir: ${ATHENA_QL}
profiling:
enabled: true
turn_off_expensive_profiling_metrics: true
include_field_distinct_count: true
include_field_min_value: true
include_field_max_value: true
include_field_mean_value: true
include_field_sample_values: true
field_sample_values_limit: 2
profile_if_updated_since_days: 10
stateful_ingestion:
enabled: true
ignore_old_state: false
ignore_new_state: false
remove_stale_metadata: true
env: PROD
pipeline_name: "BDV-prod-topdesk-transformed"
transformers: # an array of transformers applied sequentially
- type: "pattern_add_dataset_terms"
config:
term_pattern:
rules:
".*": ["urn:li:glossaryTerm:INTERN_OPEN"]
- type: simple_add_dataset_tags
config:
tag_urns:
- "urn:li:tag:Bedrijfsvoering"
- "urn:li:tag:Topdesk"
- "urn:li:tag:PROD"
- "urn:li:tag:Transformed"
- type: "simple_add_dataset_domain"
config:
replace_existing: true # false is default behaviour
domains:
- "urn:li:domain:1ef9fa01-a415-46e2-93ad-f8ce3bf84537" # domein 'Bedrijfsvoering'
adorable-forest-52600
06/21/2023, 11:25 AMlively-thailand-64294
06/21/2023, 2:58 PMrich-restaurant-61261
06/21/2023, 8:46 PMcalm-helmet-89243
06/21/2023, 10:41 PMgifted-diamond-19544
06/22/2023, 7:08 AM"Ingestion error: An error occurred (MetadataException) when calling the GetTableMetadata operation: Rate exceeded (Service: AmazonDataCatalog; Status Code: 400; Error Code: ThrottlingException
proud-dusk-671
06/22/2023, 7:41 AMcreamy-pizza-80433
06/22/2023, 10:10 AMmodern-hospital-90979
06/22/2023, 2:03 PMlooker
and lookml
ingestion patterns and they appear to be pulling in most, if not all, of our assets in the platform. However, I'm unable to locate certain specific views that are defined in Looker ad Persistent Derived Tables (PDTs). Some PDTs show up, but others do not. It's unclear if there's a pattern to which ones show up and which do not. Have other users experienced challenges ingesting Looker PDTs?strong-diamond-4751
06/22/2023, 3:41 PMgreat-notebook-53658
06/23/2023, 7:59 AMgreat-notebook-53658
06/23/2023, 8:50 AMbillions-journalist-13819
06/23/2023, 8:57 AM