colossal-balloon-71404
06/13/2023, 3:01 PMcolossal-balloon-71404
06/13/2023, 3:01 PMcolossal-balloon-71404
06/13/2023, 3:03 PMcustom_dataset.json
file:
[
{
"auditHeader":null,
"entityType":"dataset",
"entityUrn": "urn:li:dataset:(urn:li:dataPlatform:MyCustomDataPlatform,test,PROD)",
"changeType":"UPSERT",
"systemMetadata":null
}
]
colossal-balloon-71404
06/13/2023, 3:04 PMsource:
type: file
config:
# Source-type specifics config
filename: ./custom-dataset.json
Getting this result:
Source (file) report:
{'events_produced': 0,
'events_produced_per_sec': 0,
'entities': {},
'aspects': {},
'warnings': {},
'failures': {'path-0': ['com.linkedin.pegasus2avro.usage.UsageAggregation is missing required field: bucket']},
'total_num_files': 1,
'num_files_completed': 1,
'files_completed': ['custom-dataset.json'],
'percentage_completion': '0%',
'estimated_time_to_completion_in_minutes': -1,
'total_bytes_read_completed_files': 209,
'current_file_size': 209,
'total_parse_time_in_seconds': 0.0,
'total_count_time_in_seconds': 0.0,
'total_deserialize_time_in_seconds': 0,
'aspect_counts': {},
'entity_type_counts': {},
'start_time': '2023-06-13 14:37:07.777117 (now)',
'running_time': '0.77 seconds'}
Pipeline finished with at least 1 failures; produced 0 events in 0.77 seconds.
billions-baker-82097
06/13/2023, 3:04 PMcolossal-balloon-71404
06/13/2023, 3:06 PMdatahub put platform --name Zendesk --display_name "Zendesk" --logo "<https://assets.website-files.com/5a0242c3d47dd70001e5b2e9/5a054c7012148e00015864fc_zmark%401x.svg>"
It succeeded, but doesn't show up in UI anywhere. I thought it might be because it doesn't have any datasets associated with itcolossal-balloon-71404
06/13/2023, 3:07 PMbillions-baker-82097
06/13/2023, 3:09 PMcolossal-balloon-71404
06/13/2023, 3:10 PMmammoth-bear-12532
colossal-balloon-71404
06/14/2023, 6:58 AMcolossal-balloon-71404
06/14/2023, 12:13 PMmammoth-bear-12532
colossal-balloon-71404
06/15/2023, 6:19 AMcolossal-balloon-71404
06/15/2023, 6:20 AMmammoth-bear-12532
mammoth-bear-12532
colossal-balloon-71404
06/16/2023, 6:11 AMserver.database.schema.table -> raw_server.database_schema.table
It's based on our own config files for our own data loader written in Python. The idea is to run the lineage part after every time a table is loaded.
As long as we're doing it in Python in one instance it also makes sense to keep it in Python in other instances.
Anyways, it's gone far enough off topic now. Thank you for your support 🙂colossal-balloon-71404
07/10/2023, 1:38 PM