future-table-91845
05/11/2023, 5:42 PMfuture-table-91845
05/11/2023, 5:42 PMbig-postman-38407
05/12/2023, 7:36 AMbroad-ghost-1006
05/12/2023, 8:53 AMbillions-baker-82097
05/12/2023, 10:43 AMbland-orange-13353
05/12/2023, 1:16 PMsalmon-exabyte-77928
05/12/2023, 3:22 PMvalues.yaml
host: "deps-postgresql.deps.svc.cluster.local:5432"
hostForpostgresqlClient: "deps-postgresql.deps.svc.cluster.local"
port: "5432"
url: "jdbc:<postgresql://deps-postgresql.deps.svc.cluster.local:5432/dh02?verifyServerCertificate=false&useSSL=true&useUnicode=yes&characterEncoding=UTF-8&enabledTLSProtocols=TLSv1.2>"
driver: "org.postgresql.Driver"
username: "postgres"
password:
secretRef: postgresql-secrets
secretKey: postgres-password
# --------------OR----------------
# value: password
But postgresql setup job creates database datahub
and ignores database dh02
specified in connection string. Any minds?
2023/05/12 15:11:11 Waiting for: <tcp://deps-postgresql.deps.svc.cluster.local:5432>
2023/05/12 15:11:11 Connected to <tcp://deps-postgresql.deps.svc.cluster.local:5432>
CREATE DATABASE
-- create metadata aspect table
After job finished database datahub
appears on the postgresql instance and tables present there, but db dh02
is not.ancient-kitchen-28586
05/14/2023, 8:24 AMancient-kitchen-28586
05/14/2023, 8:28 AM[+] Running 0/0
- Container mysql Creating 0.0s
Error response from daemon: invalid volume specification: 'C:\Users\janko\.datahub\mysql\init.sql:/docker-entrypoint-initdb.d/init.sql:rw': invalid mount config for type "bind": bind source path does not exist: c:\users\janko\.datahub\mysql\init.sql
I don't have a mysql folder in .datahub.
Any ideas why this could be?hallowed-lock-74921
05/14/2023, 4:23 PMhallowed-lock-74921
05/14/2023, 4:23 PMhallowed-lock-74921
05/14/2023, 4:28 PMConfigure project dockermysql-setupfullVersion=v0.10.2-147-g0fa983a.dirty cliMajorVersion=0.10.2 version=0.10.3-SNAPSHOT SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". SLF4J: Defaulting to no-operation (NOP) logger implementation SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details.
Task dockermysql-setup:dockerFAILED unknown flag: --load See 'docker --help'.
hallowed-lock-74921
05/14/2023, 5:25 PMProcess 'command 'docker'' finished with non-zero exit value 125* Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== 2: Task failed with an exception. ----------- * What went wrong: Execution failed for task 'dockerelasticsearch-setup:docker'.
Process 'command 'docker'' finished with non-zero exit value 125* Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== 3: Task failed with an exception. ----------- * What went wrong: Execution failed for task 'dockermysql-setup:docker'.
Process 'command 'docker'' finished with non-zero exit value 125* Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.9.2/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 1m 10s 88 actionable tasks: 52 executed, 36 up-to-date (venv) Apples-MacBook-Pro-2:datahub apple$
proud-dusk-671
05/15/2023, 3:04 PMprehistoric-greece-5672
05/15/2023, 3:21 PMbillions-baker-82097
05/15/2023, 4:46 PMbreezy-balloon-32520
05/15/2023, 5:04 PMwitty-butcher-82399
05/16/2023, 3:45 AMThis authenticator also looks for a "delegated actor urn" which can be provided by system callers using the 'X-DataHub-Actor' header.However, the current logic does not match that, how is that? https://github.com/datahub-project/datahub/blob/master/metadata-service/auth-impl/[…]ub/authentication/authenticator/DataHubSystemAuthenticator.java
proud-dusk-671
05/16/2023, 6:06 AMbreezy-leather-30929
05/16/2023, 6:45 AMpowerful-finland-16210
05/16/2023, 6:54 AMrough-summer-14442
05/16/2023, 6:55 AMbillions-baker-82097
05/16/2023, 11:03 AMnumerous-refrigerator-15664
05/16/2023, 11:16 AMdatabase_pattern
and table_pattern
.
Additionally, I'm trying to do the following things, and I need some advice on it. Any help would be appreciated.
1. Now I'm trying to ingest most of the datasets from hive metastore
, and I was wondering if there's a way to do pattern filtering for other items in hive metastore
as well:
• DBS.DB_LOCATION_URI (e.g. allow only the pattern "hdfs://cluster1/dsc/.*")
• DBS.OWNER_NAME (e.g. deny those with accounts ".*test")
• If it is impossible via recipes, would there be any other possible ways?
2. Our organization manages additional metadata for hive's DB, table, and column in other mysql DB (say our_meta
) which is separated from hive metastore
. For example, for a table named customer.cust_mst
on the hive, this table exists in the hive metastore
, and a separate mysql DB also manages information about this table. Given the situation, I'd like to ingest the metadata of our_meta
into datahub. What should be the best way to do it?
• Some of the managed items in our_meta
(mainly technical meta) seems to be managed as custom properties, and this should be synced in batch mode.
• Some of the managed items in our_meta
(mainly business meta) can be managed as business glossary or tags, and both batch sync and API calls should possible.
• I am looking into custom ingestion source and metadata ingestion transformer, but I am not sure how to approach.
I hope I can handle these without forking the source if possible. Thanks in advance.eager-belgium-3850
05/16/2023, 1:47 PMhandsome-flower-80167
05/16/2023, 7:14 PMmost-engine-69486
05/16/2023, 7:14 PMdry-spring-48163
05/17/2023, 5:10 AMhandsome-cat-78137
05/17/2023, 9:48 AMdatahub.ingestion.run.pipeline
).
Then to update properties, tags, owners etc:
1. I search for the uploaded dataset urn using the resli api with endpoint /entities
because I do not know the urn beforehand
2. Update the urn
The problem that I face with this approach is that urn creation takes some time depending on the dataset. And usually it is not ready when I try searching for it.
Is there some way to know when the urn is created after the pipeline is run?
Or there a better way to do this?polite-cat-69516
05/18/2023, 6:11 AM