acoustic-dusk-3739
07/21/2022, 6:58 PMDeprecated
?mysterious-eye-58423
07/22/2022, 5:08 AMMetadataChangeLog
events and replaying log-compacted MetadataChangeLog
events as a mechanism to rebuild/recover search indexes particularly for time-series metadata that we don't persist in Datahub MySQL table?late-bear-87552
07/22/2022, 9:08 AMlemon-zoo-63387
07/22/2022, 9:43 AMpython3 -m datahub delete --env QA --entity_type dataset --platform hive
glamorous-library-1322
07/22/2022, 11:31 AMModuleNotFoundError
So i would like to know how can i install packages in the env this ingestion is running in.late-bear-87552
07/22/2022, 1:19 PMgentle-camera-33498
07/22/2022, 1:37 PMbrave-tomato-16287
07/22/2022, 3:03 PMfull-chef-85630
07/23/2022, 8:07 AM[root@VM-4-16-centos airflow]# airflow connections add --conn-type 'datahub_rest' 'datahub_rest_default' --conn-host '<http://localhost:8080>'
[2022-07-23 161208,690] {cli_action_loggers.py:105} WARNING - Failed to log action with (sqlite3.OperationalError) no such table: log
[SQL: INSERT INTO log (dttm, dag_id, task_id, event, execution_date, owner, extra) VALUES (?, ?, ?, ?, ?, ?, ?)]
[parameters: ('2022-07-23 081208.687171', None, None, 'cli_connections_add', None, 'root', '{"host_name": "VM-4-16-centos", "full_command": "[\'/usr/local/bin/airflow\', \'connections\', \'add\', \'--conn-type\', \'datahub_rest\', \'datahub_rest_default\', \'--conn-host\', \'http://localhost:8080\']"}')]
(Background on this error at: http://sqlalche.me/e/13/e3q8)
Traceback (most recent call last):
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/engine/base.py", line 1277, in _execute_context
cursor, statement, parameters, context
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/engine/default.py", line 608, in do_execute
cursor.execute(statement, parameters)
sqlite3.OperationalError: no such table: connection
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/bin/airflow", line 8, in <module>
sys.exit(main())
File "/usr/local/lib/python3.6/site-packages/airflow/__main__.py", line 48, in main
args.func(args)
File "/usr/local/lib/python3.6/site-packages/airflow/cli/cli_parser.py", line 48, in command
return func(*args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/airflow/utils/cli.py", line 92, in wrapper
return f(*args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/airflow/cli/commands/connection_command.py", line 196, in connections_add
if not session.query(Connection).filter(Connection.conn_id == new_conn.conn_id).first():
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/orm/query.py", line 3429, in first
ret = list(self[0:1])
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/orm/query.py", line 3203, in getitem
return list(res)
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/orm/query.py", line 3535, in iter
return self._execute_and_instances(context)
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/orm/query.py", line 3560, in _execute_and_instances
result = conn.execute(querycontext.statement, self._params)
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/engine/base.py", line 1011, in execute
return meth(self, multiparams, params)
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/sql/elements.py", line 298, in _execute_on_connection
return connection._execute_clauseelement(self, multiparams, params)
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/engine/base.py", line 1130, in _execute_clauseelement
distilled_params,
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/engine/base.py", line 1317, in _execute_context
e, statement, parameters, cursor, context
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/engine/base.py", line 1511, in _handle_dbapi_exception
sqlalchemy_exception, with_traceback=exc_info[2], from_=e
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/util/compat.py", line 182, in raise_
raise exception
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/engine/base.py", line 1277, in _execute_context
cursor, statement, parameters, context
File "/usr/local/lib64/python3.6/site-packages/sqlalchemy/engine/default.py", line 608, in do_execute
cursor.execute(statement, parameters)
sqlalchemy.exc.OperationalError: (sqlite3.OperationalError) no such table: connection
[SQL: SELECT connection.password AS connection_password, connection.extra AS connection_extra, connection.id AS connection_id, connection.conn_id AS connection_conn_id, connection.conn_type AS connection_conn_type, connection.description AS connection_description, connection.host AS connection_host, connection.schema AS connection_schema, connection.login AS connection_login, connection.port AS connection_port, connection.is_encrypted AS connection_is_encrypted, connection.is_extra_encrypted AS connection_is_extra_encrypted
FROM connection
WHERE connection.conn_id = ?
LIMIT ? OFFSET ?]
[parameters: ('datahub_rest_default', 1, 0)]
(Background on this error at: http://sqlalche.me/e/13/e3q8)lemon-zoo-63387
07/25/2022, 12:22 AMlemon-zoo-63387
07/25/2022, 12:51 AMbitter-tent-3827
07/25/2022, 5:58 AMsquare-hair-99480
07/25/2022, 6:51 AMprofiling.profile_table_size_limit
& profiling.profile_table_row_limit
also for Snowflake? ... I see now it works only for BigQuerycool-vr-73109
07/25/2022, 6:53 AMsquare-hair-99480
07/25/2022, 9:49 AMaccess_history
but when I check this view in my Snowflake enterprise account it is empty and hence I have no lineage appearing in Datahub. Anyone has faced something similar?gentle-camera-33498
07/25/2022, 2:33 PMcolossal-sandwich-50049
07/25/2022, 3:50 PMnice-country-99675
07/25/2022, 5:47 PMdazzling-insurance-83303
07/26/2022, 3:49 AMallow_deny_pattern
signify?
profiling:
enabled: true
allow_deny_patterns:
allow:
- .*
deny:
-
ignoreCase: True
alphabet: '[A-Za-z0-9 .-]'
Is that filtering for data within the columns? If so, are there any examples to refer to?
I am interested in knowing if those can be regexes to do Luhn algorithm checks.able-evening-90828
07/26/2022, 5:08 AMcd metadata-ingestion
../gradlew :metadata-ingestion:installDev
source venv/bin/activate
Then I tried to ingest something from mysql using the command below
python3 -m datahub ingest -c ../test.mysql.localhost.dhub.yml
And I got the following mysterious error.
Failed to create source due to mysql is disabled due to an error in initialization
Some small instrumentation of code revealed the exception to be
dlopen(/Users/jinlin/Code/datahub/metadata-ingestion/venv/lib/python3.9/site-packages/greenlet/_greenlet.cpython-39-darwin.so, 0x0002): tried: '/Users/jinlin/Code/datahub/metadata-ingestion/venv/lib/python3.9/site-packages/greenlet/_greenlet.cpython-39-darwin.so' (mach-o file, but is an incompatible architecture (have 'x86_64', need 'arm64e'))
I am on a Mac with M1 chip and this looks like a mismatch between M1 binary and x86 binary. What should I do to make this working?cool-vr-73109
07/26/2022, 9:26 AMlemon-terabyte-66903
07/26/2022, 3:04 PMadamant-mouse-7290
07/26/2022, 4:10 PMtype: athena
config:
aws_region: xxx
work_group: xxx
username: '${xxx}'
password: '${xxx}'
s3_staging_dir: 'xxx'
include_views: true
include_tables: true
database: table1, table2 ... table10
sink:
type: datahub-rest
config:
server: 'xxx'
token: xxx
busy-analyst-8258
07/26/2022, 7:24 PMkind-whale-32412
07/26/2022, 11:49 PM] ERROR {datahub.ingestion.run.pipeline:273} - Failed to extract some records due to: source produced an invalid metadata work unit: MetadataChangeEventClass...
cuddly-arm-8412
07/27/2022, 3:30 AMwooden-chef-22394
07/27/2022, 6:31 AMFailed to create source due to clickhouse is disabled; try running: pip install 'acryl-datahub[clickhouse]'
gifted-knife-16120
07/27/2022, 7:33 AMgifted-knife-16120
07/27/2022, 7:32 AMcontainer
but, i get this error
my command is : datahub delete --urn "urn:li:container:(13c86013c4ae5a2027b9e2f2b9443a91)" --soft
faint-advantage-18690
07/27/2022, 7:49 AM