few-air-56117
01/04/2022, 12:13 PMgentle-nest-904
01/04/2022, 12:26 PMgentle-nest-904
01/04/2022, 12:27 PMnice-country-99675
01/04/2022, 2:22 PMreadhisft-usage
new source ingestor.... and I found this scenario. In the regular redshift
metadata ingestor I have a database alias, and some schemas and tables filtered out. Is there a way to apply the same settings the the redshift usage
recipe?few-air-56117
01/04/2022, 5:49 PMsome-crayon-90964
01/04/2022, 9:52 PMnice-planet-17111
01/05/2022, 1:15 AMdbt
on bigquery... until i realized ingesting bigquery now supports lineage information. So, I was testing out to use lineage information only out of bigquery, after hard deleting all dbt entities. However, When i do it.. dbt entities(=nodes) still appear in the lineage graphs like haunting ghosts .. How can i REALLY delete them ? đloud-holiday-22352
01/05/2022, 7:10 AMgentle-sundown-2310
01/05/2022, 6:38 PMsource:
type: mysql
config:
# Coordinates
host_port:
database: tableau_advocate_lsnr
# Credentials
username:
password:
schema_pattern:
allow:
- "tableau_advocate_lsnr"
profiling:
enabled: true
profile_pattern:
allow:
- ".*standard_module_daily"
sink:
type: "datahub-rest"
config:
server: "<http://localhost:8080>"
gentle-sundown-2310
01/05/2022, 6:39 PM2 validation errors for MySQLConfig
profile_pattern
extra fields not permitted (type=value_error.extra)
profiling
extra fields not permitted (type=value_error.extra)
wide-helicopter-97009
01/05/2022, 9:42 PMdamp-ambulance-34232
01/06/2022, 4:10 AMgentle-florist-49869
01/06/2022, 8:33 PMbumpy-translator-90745
01/06/2022, 8:40 PMpip install 'acryl-datahub[bigquery]'
. I am getting the following error. I am not sure how to resolve it. Thanks!gray-wall-52477
01/06/2022, 11:59 PMsource:
type: mysql
config:
host_port: XXX
database: XXX
username: XXX
password: XXX
env: Production
include_views: False
sink:
type: "datahub-rest"
config:
server: "<http://localhost:8080>"
It runs fine and I get this as output:
{'entities_profiled': 0,
'failures': {},
'filtered': [],
'query_combiner': None,
'soft_deleted_stale_entities': [],
'tables_scanned': 215,
'views_scanned': 0,
'warnings': {'DB.AAA': ['unable to map type BIT(length=1) to metadata schema'],
'DB.BBB': ['unable to map type BIT(length=1) to metadata schema'],
'DB.CCC': ['unable to map type BIT(length=1) to metadata schema'],
'DB.DDD': ['unable to map type BIT(length=1) to metadata schema']},
'workunit_ids': [LIST OF TABLES HERE],
'workunits_produced': 215}
Sink (datahub-rest) report:
{'downstream_end_time': None,
'downstream_start_time': None,
'downstream_total_latency_in_seconds': None,
'failures': [],
'records_written': 0,
'warnings': []}
As you can see, it says 215 tables are scanned and there are 4 warnings but records_written
is 0 đ¤damp-ambulance-34232
01/07/2022, 2:38 AM[2022-01-07 09:36:48,326] ERROR {datahub.ingestion.run.pipeline:86} - failed to write record with workunit hive.ghtk.table_with_struct_type with ('Unable to emit metadata to DataHub GMS', {'exceptionClass': 'com.linkedin.restli.server.RestLiServiceException', 'stackTrace': 'com.linkedin.restli.server.RestLiServiceException [HTTP Status:422]: com.linkedin.metadata.entity.ValidationException: Failed to validate record with class com.linkedin.entity.Entity: ERROR :: /value/com.linkedin.metadata.snapshot.DatasetSnapshot/aspects/2/com.linkedin.schema.SchemaMetadata/fields/9/jsonProps :: unrecognized field found but not allowed\nERROR :: /value/com.linkedin.metadata.snapshot.DatasetSnapshot/aspects/2/com.linkedin.schema.SchemaMetadata/fields/10/jsonProps :: unrecognized field found but not allowed\nERROR ::
red-piano-51229
01/07/2022, 8:13 AMquaint-branch-37931
01/07/2022, 3:17 PMgentle-florist-49869
01/07/2022, 3:19 PMquaint-branch-37931
01/10/2022, 12:56 PMclever-australia-61035
01/10/2022, 2:32 PMwide-helicopter-97009
01/10/2022, 7:55 PMshy-parrot-64120
01/11/2022, 12:25 AMkafka-connect
extra for ingestion?
receiving following error:
File "/home/airflow/.local/lib/python3.9/site-packages/jpype/_jvmfinder.py", line 212, in get_jvm_path
raise JVMNotFoundException("No JVM shared library file ({0}) "
jpype._jvmfinder.JVMNotFoundException: No JVM shared library file (libjvm.so) found. Try setting up the JAVA_HOME environment variable properly.
salmon-rose-54694
01/11/2022, 2:11 AM[
{
"auditHeader": null,
"proposedSnapshot": {
"com.linkedin.pegasus2avro.metadata.snapshot.DatasetSnapshot": {
"urn": "urn:li:dataset:(urn:li:dataPlatform:mysql,abtest.abtest.abtestv3_allocation,PROD)",
"aspects": [
{
"com.linkedin.pegasus2avro.schema.SchemaMetadata": {
"schemaName": "abtest.abtest.abtestv3_allocation",
"platform": "urn:li:dataPlatform:mysql",
"version": 1,
"created": {
"time": 0,
"actor": "urn:li:corpuser:unknown",
"impersonator": null
},
"lastModified": {
"time": 1641798176000,
"actor": "urn:li:corpuser:unknown",
"impersonator": null
},
"deleted": null,
"dataset": null,
"cluster": null,
"hash": "",
"platformSchema": {
"com.linkedin.pegasus2avro.schema.MySqlDDL": {
"tableSchema": ""
}
},
"fields": [
{
"fieldPath": "id",
"jsonPath": null,
"nullable": false,
"description": null,
"type": {
"type": {
"com.linkedin.pegasus2avro.schema.NumberType": {}
}
},
"nativeDataType": "INTEGER(display_width=11)",
"recursive": false,
"globalTags": null,
"glossaryTerms": null
}
],
"primaryKeys": null,
"foreignKeysSpecs": null
}
}
]
}
},
"proposedDelta": null,
"systemMetadata": {
"lastObserved": 1629696884482,
"runId": "d2584674-03d3-11ec-8de4-9ae590158f91",
"properties": null
}
}
]
melodic-helmet-78607
01/11/2022, 3:54 AMthankful-businessperson-69424
01/11/2022, 12:16 PMsome-crayon-90964
01/11/2022, 4:14 PMgentle-florist-49869
01/11/2022, 6:15 PMwide-helicopter-97009
01/11/2022, 8:46 PMred-pizza-28006
01/12/2022, 1:06 PM[2022-01-12 14:03:58,671] ERROR {datahub.ingestion.run.pipeline:85} - failed to write record with workunit operation-aspect-SUMUP_DWH_PROD.ACCESS_MANAGER.ACCESS_MANAGER_REVOKE_LIST-2022-01-11T23:34:10.293000+00:00 with ('Unable to emit metadata to DataHub GMS', {'exceptionClass': 'com.linkedin.restli.server.RestLiServiceException', 'stackTrace': 'com.linkedin.restli.server.RestLiServiceException [HTTP Status:500]: java.lang.RuntimeException: Unknown aspect operation for entity dataset\n\tat com.linkedin.metadata.restli.RestliUtil.toTask(RestliUtil.java:42)\n\tat com.linkedin.metadata.restli.RestliUtil.toTask(RestliUtil.java:50)\n\tat com.linkedin.metadata.resources.entity.AspectResource.ingestProposal(AspectResource.java:132)\n\tat sun.reflect.GeneratedMethodAccessor245.invoke(Unknown Source)\n\tat sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)\n\tat java.lang.reflect.Method.invoke(Method.java:498)\n\tat com.linkedin.restli.internal.server.RestLiMethodInvoker.doInvoke(RestLiMethodInvoker.java:172)\n\tat com.linkedin.restli.internal.server.RestLiMethodInvoker.invoke(RestLiMethodInvoker.java:326)\n\tat com.linkedin.restli.internal.server.filter.FilterChainDispatcherImpl.onRequestSuccess(FilterChainDispatcherImpl.java:47)\n\tat com.linkedin.restli.internal.server.filter.RestLiFilterChainIterator.onRequest(RestLiFilterChainIterator.java:86)\n\tat com.linkedin.restli.internal.server.filter.RestLiFilterChainIterator.lambda$onRequest$0(RestLiFilterChainIterator.java:73)\n\tat java.util.concurrent.CompletableFuture.uniAccept(CompletableFuture.java:670)\n\tat java.util.concurrent.CompletableFuture.uniAcceptStage(CompletableFuture.java:683)\n\tat java.util.concurrent.CompletableFuture.thenAccept(CompletableFuture.java:2010)\n\tat com.linkedin.restli.internal.server.filter.RestLiFilterChainIterator.onRequest(RestLiFilterChainIterator.java:72)\n\tat com.linkedin.restli.internal.server.filter.RestLiFilterChain.onRequest(RestLiFilterChain.java:55)\n\tat com.linkedin.restli.server.BaseRestLiServer.handleResourceRequest(BaseRestLiServer.java:218)\n\tat com.linkedin.restli.server.RestRestLiServer.handleResourceRequestWithRestLiResponse(RestRestLiServer.java:242)\n\tat com.linkedin.restli.server.RestRestLiServer.handleResourceRequest(RestRestLiServer.java:211)\n\tat com.linkedin.restli.server.RestRestLiServer.handleResourceRequest(RestRestLiServer.java:181)\n\tat com.linkedin.restli.server.RestRestLiServer.doHandleRequest(RestRestLiServer.java:164)\n\tat com.linkedin.restli.server.RestRestLiServer.handleRequest(RestRestLiServer.java:120)\n\tat com.linkedin.restli.server.RestLiServer.handleRequest(RestLiServer.java:132)\n\tat com.linkedin.restli.server.DelegatingTransportDispatcher.handleRestRequest(DelegatingTransportDispatcher.java:70)\n\tat com.linkedin.r2.filter.transport.DispatcherRequestFilter.onRestRequest(DispatcherRequestFilter.java:70)\n\tat com.linkedin.r2.filter.TimedRestFilter.onRestRequest(TimedRestFilter.java:72)\n\tat com.linkedin.r2.filter.FilterChainIterator$FilterChainRestIterator.doOnRequest(FilterChainIterator.java:146)\n\tat com.linkedin.r2.filter.FilterChainIterator$FilterChainRestIterator.doOnRequest(FilterChainIterator.java:132)\n\tat com.linkedin.r2.filter.FilterChainIterator.onRequest(FilterChainIterator.java:62)\n\tat com.linkedin.r2.filter.TimedNextFilter.onRequest(TimedNextFilter.java:55)\n\tat com.linkedin.r2.filter.transport.ServerQueryTunnelFilter.onRestRequest(ServerQueryTunnelFilter.java:58)\n\tat com.linkedin.r2.filter.TimedRestFilter.onRestRequest(TimedRestFilter.java:72)\n\tat com.linkedin.r2.filter.FilterChainIterator$FilterChainRestIterator.doOnRequest(FilterChainIterator.java:146)\n\tat com.linkedin.r2.filter.FilterChainIterator$FilterChainRestIterator.doOnRequest(FilterChainIterator.java:132)\n\tat com.linkedin.r2.filter.FilterChainIterator.onRequest(FilterChainIterator.java:62)\n\tat com.linkedin.r2.filter.TimedNextFilter.onRequest(TimedNextFilter.java:55)\n\tat com.linkedin.r2.filter.message.rest.RestFilter.onRestRequest(RestFilter.java:50)\n\tat com.linkedin.r2.filter.TimedRestFilter.onRestRequest(TimedRestFilter.java:72)\n\tat com.linkedin.r2.filter.FilterChainIterator$FilterChainRestIterator.doOnRequest(FilterChainIterator.java:146)\n\tat com.linkedin.r2.filter.FilterChainIterator$FilterChainRestIterator.doOnRequest(FilterChainIterator.java:132)\n\tat com.linkedin.r2.filter.FilterChainIterator.onRequest(FilterChainIterator.java:62)\n\tat com.linkedin.r2.filter.FilterChainImpl.onRestRequest(FilterChainImpl.java:96)\n\tat com.linkedin.r2.filter.transport.FilterChainDispatcher.handleRestRequest(FilterChainDispatcher.java:75)\n\tat com.linkedin.r2.util.finalizer.RequestFinalizerDispatcher.handleRestRequest(RequestFinalizerDispatcher.java:61)\n\tat com.linkedin.r2.transport.http.server.HttpDispatcher.handleRequest(HttpDispatcher.java:101)\n\tat <http://com.linkedin.r2.transport.ht|com.linkedin.r2.transport.ht>
Any ideas?