https://datahubproject.io logo
Join Slack
Powered by
# getting-started
  • r

    refined-gold-30439

    06/19/2023, 7:52 AM
    Hi, I am currently using datahub quicksight version 0.10.4.0 for my PoC. However, I keep getting the error "Unable to login (invalid/expired credentials or missing permissions)" when trying to ingest Tableau metadata. Is this an expected issue since Tableau Server is an installed version and not cloud? Is there anything else I should be looking at to troubleshoot? Test Condition: - Tableau Server version: 2021.4.5 (20214.22.0314.1536) 64-bit Linux - Enable the Tableau Metadata API - Tableau Credentials (Username/Password or Personal Access Token) - Site Administrator Explorer permission
    d
    • 2
    • 2
  • b

    bland-orange-13353

    06/19/2023, 9:28 AM
    This message was deleted.
    ✅ 1
    j
    • 2
    • 1
  • d

    damp-student-67816

    06/19/2023, 11:33 AM
    Hi team, I have integrated oidc with my datahub self hosted site. It is working as expected. I want to know where these user details are stored in datahub. Also where I can view these user details from code or db Not from UI?
    ✅ 1
    d
    • 2
    • 1
  • j

    jolly-airline-17196

    06/20/2023, 8:29 AM
    Hey, on the demo page, we have a HDFS platform. how do we connect HDFS to the datahub, what is the ingestion procedure, someone provide me with the recipe
    d
    f
    • 3
    • 2
  • b

    blue-state-10999

    06/20/2023, 11:29 AM
    Hi, we are currently running datahub in our kubernetes cluster and were wondering if it is possible to run gms and the frontend components with multiple replicas?
    d
    a
    • 3
    • 3
  • l

    loud-advantage-97409

    06/20/2023, 12:50 PM
    Hi, is it possible that Datahub Helm Chart contains errors ? If Datahub values metadata_service_authentication is enabled, the datahub-datahub-system-update-job is deployed with env attributes which are relevant for metadata_service_authentication. - name: DATAHUB_SYSTEM_CLIENT_ID value: __datahub_system - name: DATAHUB_SYSTEM_CLIENT_SECRET valueFrom: secretKeyRef: key: system_client_secret name: datahub-auth-secrets the service is deployed and started before secret (datahub-auth-secrets) is created. datahub-datahub-system-update-job use helm hook pre-install, pre-upgrade. Secret datahub-auth-secrets.yaml was defined without helm hook, so will be executed afterwards.
    d
    c
    • 3
    • 2
  • c

    calm-agency-11588

    06/20/2023, 6:18 PM
    Hi - I am running the datahub quickstart version on my local. The default host name for my localhost is picked as 0.0.0.0 for elasticsearch. How can I change it to 127.0.0.1 in my docker-compose.yml file ?
    ✅ 1
    d
    • 2
    • 1
  • s

    sparse-photographer-68618

    06/21/2023, 3:26 AM
    Hi team, I am interested in re-build datahub components image with Dockerfile in your official github, but it seems those Dockerfile cannot find some .jar or .war file for COPY task, is there any extra step to use those Dockerfile
    ✅ 1
    b
    • 2
    • 4
  • m

    magnificent-plumber-63682

    06/21/2023, 11:21 AM
    I am trying to build datahub metadata service image locally using ./gradlew metadata servicewar:build. But getting an error
    Task li utilsgenerateDataTemplate FAILED
    There are 34 data schema input files. Using input root folder: D:\Nuodata\datahub\li-utils\src\main\pegasus FAILURE: Build failed with an exception. * What went wrong: Execution failed for task 'li utilsgenerateDataTemplate'.
    'other' has different root
    Can some please help me here?
    ✅ 1
    d
    • 2
    • 3
  • q

    quick-teacher-37159

    06/21/2023, 2:51 PM
    Hi team, I want to know if we can update the metadata of a business glossary via datahub APIs? I was trying the Python Emitter and it worked well with the the datasets documentation updates, I need to replicate the same in Business Glossary object.
    b
    l
    d
    • 4
    • 6
  • b

    billions-oxygen-40386

    06/22/2023, 5:51 AM
    Hello team, our organisation is currently in the process of implementing Datahub as our data-catalog tool. In order to successfully complete our security assessment, we require a recent Soc2/ISO27001 audit report. Would it be possible for anyone of you to kindly guide me in the right direction to obtain this report?
    ✅ 1
    d
    • 2
    • 1
  • b

    bitter-translator-92563

    06/22/2023, 8:55 PM
    Hi guys. At the moment we are working on few new features in DataHub which will require some changes in UI. So we got a question - is there a design system for DataHub? So we can make UI changes look more organic
    d
    f
    • 3
    • 3
  • p

    proud-dusk-671

    06/26/2023, 7:34 AM
    Hi, Trying to use APIs in Datahub for the first time. I get the following error on executing the query -
    Copy code
    curl -v --user <admin username>:<admin password> --location --request POST 'http://<domain url>/api/graphql' \
    --header 'Content-Type: application/json' \
    --data-raw '{ "query": "mutation createDomain { createDomain(input: { name: \"lol\", description: \"nothing\" }) }", "variables":{}}'
    Can you tell me what I am missing? Error
    Copy code
    * Mark bundle as not supporting multiuse
    < HTTP/1.1 401 Unauthorized
    < Date: Mon, 26 Jun 2023 07:33:08 GMT
    d
    • 2
    • 2
  • f

    fancy-toddler-89669

    06/26/2023, 12:48 PM
    Are there any definitions specified on the metrics within the analytics page? We have a few question on them which we can't find answers for. For example are the weekly monthly active users unique? Or will I be counted twice if I for log in on both Monday and Tuesday?
    ✅ 1
    d
    b
    • 3
    • 4
  • s

    square-jelly-83571

    06/26/2023, 4:00 PM
    Hello team, I want to ingest csv file to datahub. I was trying to configure the recipe: https://datahubproject.io/docs/generated/ingestion/sources/csv/ Here I want to use the URL of file. So can someone tell what kind of URL? I have 3 csv files
    ✅ 1
    d
    • 2
    • 1
  • r

    rich-restaurant-61261

    06/26/2023, 10:58 PM
    Hi Team, I saw in the Feb 2023 town hall, it shows a really cool and useful feature -- notification, but my setting page don't have that notification tab, anyone know if the feature has been release or not? the values version I use is v0.10.4.
    ✅ 1
    d
    d
    l
    • 4
    • 5
  • p

    proud-dusk-671

    06/27/2023, 8:53 AM
    Hi, I am looking to find out how to keep different instances of Datahub in sync (non-prod and prod). We are running two instances of Datahub. The non-prod instance is obviously ahead of the prod instance. However, since the entities and datasets on both the instances are just the same, how can we keep the two in sync? For example, we do not want to be adding description on some table A on the non-prod Datahub and then after 3 months when table A is released in production, go on to manually add the description again on the prod Datahub
    d
    d
    • 3
    • 2
  • v

    victorious-gigabyte-26271

    06/27/2023, 10:40 AM
    Hi Team, I'm looking for some inputs on datahub capabilities to extract and harvest the lineage from Avro nested schema. Airflow will be the orchestrator in this case and I'd like to understand will I be able to extract the table level, column level and computation level granularity from the extracted lineage to establish data observability and traceability ?
    d
    b
    d
    • 4
    • 3
  • s

    some-furniture-28978

    06/28/2023, 7:56 AM
    I backed up the data before nuke, and then I imported it using the restore command, but datahub's ui didn't see any data,how can i do
    d
    • 2
    • 2
  • r

    ripe-eye-60209

    06/28/2023, 12:23 PM
    Hello Team, can this kafka sink (datahub-kafka) be used without schema_registry_url ? since, in recent datahub versions schema registry was removed.
    ✅ 1
    d
    b
    • 3
    • 2
  • n

    nutritious-lighter-88459

    06/28/2023, 3:39 PM
    Hi, I was experimenting with
    datahub cli
    and wanted to understand if we can create Domains and Tags with it?
    ✅ 1
    d
    • 2
    • 1
  • m

    many-toothbrush-93388

    06/29/2023, 7:39 AM
    Good morning you guys, I am a working student and I'm currently in my first Data Science project. For most of us, its our first 'real' Project, so most of my team lacks experience in some areas. 🙂 We're working with Azure and are storing our data in a Parquet and CSV format. I am tasked with researching and implementing the possibilities of a data catalog and have been recommended the Datahub project. I was wondering if it's possible to integrate datahub into our Pipeline, given our infrastructure. When going through the documentation I sadly couldn't find much about how to work with parquet data or integrating it with Azure. I would be super thankful if anyone could link me a resource for our use case so that I can get started. Thank you very much in advance ! :)
    d
    a
    • 3
    • 3
  • p

    proud-dusk-671

    06/29/2023, 8:38 AM
    Hi team, Can you tell me what is difference between
    type: INTERNAL
    and
    type:KAFKA
    in cp-schema registry and which one to be used under what circumstances?
    d
    • 2
    • 1
  • b

    broad-parrot-31743

    06/30/2023, 3:31 AM
    Hi team, does SSO support CAS?
    d
    • 2
    • 2
  • b

    broad-parrot-31743

    07/03/2023, 2:26 AM
    hi team! How to solve the problem in datahub: elasticsearch unauthorized access vulnerability?
    ✅ 1
    b
    • 2
    • 4
  • n

    nutritious-photographer-79168

    07/03/2023, 12:12 PM
    Hi team. I have recently deployed datahub to onsite kubernetes using helm chart and now i would like to try and integrate with AD using custom jaas.conf file. How do i "mount" the custom jaas.conf file to k8s pod deployed using helm? I know i need to setup a volume and volume mount. I just have no idea how.
    ✅ 1
    d
    • 2
    • 2
  • b

    bumpy-architect-30714

    07/03/2023, 4:26 PM
    Hi team, We're currently considering using datahub. I have played with the demo and searched through the docs (mostly the
    Dataset
    docs) but I can't find much regarding dataset versioning. 1. Is the same dataset mapped across different environments? E.g. if I have a dataset in dev and prod, are they automatically linked? Or are they treated as entirely different datasets because the URN is different? 2. Is there a versioning concept that allows to track datasets across multiple versions? It looks like schema changes are tracked, but can I tag those versions and use those tags to jump back in time to look at the definition of a table? 3. If it is possible to track versions: Is it possible to track a dataset when its table name changed? In major versions we might sometimes rename tables or move them to more fitting schemas. I've read somewhere that only non-breaking changes are supported, would a table rename be considered a breaking change?
    ✅ 1
    m
    m
    • 3
    • 2
  • b

    brave-judge-32701

    07/04/2023, 2:22 AM
    when I try ./gradlew quickstart or ./gradlew quickstartDebug thr kafka broker container start faild with : The Cluster ID z2hW0NRNSXG-U5P1ksnw_Q doesn’t match stored clusterId Some(Bptu0FjkQDOQqD85wd8law) in meta.properties. The broker is trying to join the wrong cluster. Configured zookeeper.connect may be wrong. How can I fix this problem.
    ✅ 1
    h
    • 2
    • 3
  • n

    nutritious-lighter-88459

    07/04/2023, 12:17 PM
    Hi team, Is there support to create assertions via GraphQL apis? I see that there is a query
    assertion(urn: String!)
    to fetch existing assertion and a mutation
    deleteAssertion(urn: String!)
    to delete assertion. But couldn't find a mutation to create a new one. Is it part of some other mutation ? TIA
    ✅ 1
    d
    a
    b
    • 4
    • 14
  • r

    rich-crowd-33361

    07/05/2023, 6:40 PM
    Hi Maggie, Can we prioritize column level lineage for Snowflake between View to view? This is one of the pain point we have in our organization to use Datahub
    ✅ 1
    d
    a
    h
    • 4
    • 3
1...676869...80Latest