https://datahubproject.io logo
Join SlackCommunities
Powered by
# troubleshoot
  • g

    gentle-camera-33498

    06/22/2022, 12:47 PM
    Hello, I'm testing the editing and glossary creation functionalities in the UI and I got the behavior below when trying to add a documentation. Would this be a bug? Datahub version: v0.8.38 Deployment type: Without neo4j and on a GKE cluster.
    b
    • 2
    • 9
  • g

    glamorous-easter-15952

    06/22/2022, 5:26 PM
    If I want to install a new plugin for sources in an existing datahub instance, which docker container should I install into?
    e
    • 2
    • 8
  • p

    prehistoric-knife-90526

    06/22/2022, 6:13 PM
    Hi All ๐Ÿ‘‹. Does anyone know what it means to get a 404 on
    aspects?action=ingestProposal
    when trying to delete a dataset using the CLI?
    e
    o
    • 3
    • 17
  • s

    salmon-angle-92685

    06/23/2022, 8:30 AM
    Hello everyone, Since yesterday, for some reason I still don't understand, all ingested data stopped showing up on the frontend http://localhost:9002/ I'm deploying the DataHub locally with quickstart. Yesterday morning all ingestions were working perfectly. And since I'm still developing the DataHub ingestion pipelines, I restart (with docker nuke and docker quickstart) the container frequently to test if my scripts work. But all of a sudden the ingestions stopped showing up on the frontend. In the terminal, there is no error. All ingestions end successfully. I checked the DataHub database and all data is correctly stored. I figured it could be the Elasticsearch indices that for some reason weren't being updated by the page, so I launched
    Copy code
    ./docker/datahub-upgrade/datahub-upgrade.sh -u RestoreIndices
    to reset the indices... But nothing worked. I tried ingesting the data directly into the UI and nothing happens. Not even the card with the name of the ingestion appears. Which makes me imagine that there is some communication problem between the interface and the bank. I asked my colleague to git pull my script and run it on his computer, and it worked. Can anybody help me ? I do not know what else to do. Thanks a lot!
    o
    • 2
    • 1
  • b

    bland-easter-53873

    06/23/2022, 8:48 AM
    Hi, have a quick question. For the metadata ingestion for snowflake, should the user have any specific role. I have full access on the schema but it is throwing login error, which is working absolutely fine when used with other applications
    s
    • 2
    • 7
  • a

    acoustic-carpenter-87261

    06/23/2022, 12:27 PM
    Hi, I am facing this issue for some of my dags https://github.com/datahub-project/datahub/issues/4546 . Is there any workaround for this?
    o
    d
    r
    • 4
    • 3
  • c

    chilly-elephant-51826

    06/23/2022, 2:09 PM
    @big-carpet-38439 need help in understanding datahub actions code what I am able to figure out is that when an ingestion is ran it executes and pipeline in datahub actions that triggers executor_actions which gets the config and run the act method I am unable to get two things 1. how the below code is able to get the task config
    Copy code
    # Build default task config
            local_task_config = TaskConfig(
                name="RUN_INGEST",
                type="acryl.executor.execution.sub_process_ingestion_task.SubProcessIngestionTask",
                configs=dict({}),
            )
    2. How act method calls the library in metadata-ingestion module ? please let me know in case my understanding is wrong or I am missing something
    o
    • 2
    • 6
  • p

    plain-farmer-27314

    06/23/2022, 2:22 PM
    Hey all! We are trying to determine who exactly our WAU's are. Is this info currently stored anywhere? Where are most of the stats in the default analytics panel stored? Thanks!
    o
    l
    • 3
    • 5
  • h

    handsome-football-66174

    06/23/2022, 4:44 PM
    Hi Everyone, Trying to use https://datahubproject.io/docs/metadata-ingestion/transformers/#mark-dataset-status But don't see the change reflect (the dataset is still visible in the UI)
    Copy code
    {
      "source": {
        "type": "glue",
        "config": {
          "env": "PROD",
          "aws_region": "us-east-1",
          "extract_transforms": "false",
          "table_pattern": {
            "allow": [
              "testdb.*$"
            ],
            "ignoreCase": "false"
          }
        }
      },
      "transformers": [
        {
          "type": "mark_dataset_status",
          "config": {
            "removed": "true"
          }
        }
      ]
    }
    o
    • 2
    • 6
  • b

    bland-barista-59197

    06/23/2022, 5:11 PM
    Hi all, Please help me to get all dataset base on browspath startwith using GraphiQL. Here is my GraphiQL
    Copy code
    {
      search(
        input: {type: DATASET, query: "*", start: 0, count: 10000, 
          filters: [{field: "browsePaths", value: "/prod/myDept/sakilaDb"} ]}
      ) {
        start
        count
        total
        searchResults {
          entity {
            urn
            type
            ... on Dataset {
              name
            }
          }
        }
      }
    }
    e
    o
    • 3
    • 3
  • h

    helpful-painting-48754

    06/27/2022, 8:56 AM
    Hi, I got this error while ingesting from my database.
    Copy code
    " 'warnings': {'cardcore_issuing_2.view_cardcore_card_master': ['Ingestion error: Table or view named '\n"
               "                                                               '`cardcore_issuing_2`.`view_cardcore_card_master` could not be reflected: "
               "'\n"
               '                                                               \'(pymysql.err.OperationalError) (1356, "View \'\n'
               '                                                               "\'cardcore_issuing_2.view_cardcore_card_master\' references invalid '
               'table(s) or "\n'
               "                                                               'column(s) or function(s) or definer/invoker of view lack rights to use "
               'them")\\n\'\n'
               "                                                               '[SQL: DESCRIBE `cardcore_issuing_2`.`view_cardcore_card_master`]\\n'\n"
               "                                                               '(Background on this error at: <http://sqlalche.me/e/13/e3q8)>'],\n"
    May I know what is the cause of this error?
    d
    • 2
    • 2
  • s

    steep-furniture-57251

    06/27/2022, 10:13 AM
    Hello everyone, I installed datahub cli and ingested data from big query, i. want to extract the tags available in a dataset. For doing so, I added a tag on dataset from the UI. and used below api to fetch the tags and other aspects of the dataset but it shows
    Copy code
    error: [HTTP Status:404]: No root resource defined for path '/prod'
    whereas when I try to search the tag , it gets correct result. urn that I am using is this : urnlidataset:(urnlidataPlatform:bigquery,temp,PROD) can anyone help me, point out what am I missing here, seems like i need to set some environment variable to point it to correct directory but i am not sure which directory it should point to.
    h
    • 2
    • 5
  • w

    wide-apple-53149

    06/27/2022, 11:25 AM
    DOCKER_BUILDKIT=1 docker build -t your_datahub_frontend -f ./docker/datahub-frontend/Dockerfile
  • w

    wide-apple-53149

    06/27/2022, 11:25 AM
    #16 160.3 * What went wrong: #16 160.3 A problem occurred configuring project ':datahub-frontend'. #16 160.3 > Could not resolve all artifacts for configuration 'datahub frontendclasspath'. #16 160.3 > Could not resolve io.github.kobylynskyigraphql java codegen gradle plugin4.1.1. #16 160.3 Required by: #16 160.3 project :datahub-frontend > io.github.kobylynskyi.graphql.codegenio.github.kobylynskyi.graphql.codegen.gradle.plugin4.1.1 #16 160.3 > Could not resolve io.github.kobylynskyigraphql java codegen gradle plugin4.1.1. #16 160.3 > Could not get resource 'https://plugins.gradle.org/m2/io/github/kobylynskyi/graphql-java-codegen-gradle-plugin/4.1.1/graphql-java-codegen-gradle-plugin-4.1.1.pom'. #16 160.3 > Could not GET 'https://plugins-artifacts.gradle.org/io.github.kobylynskyi/graphql-java-codegen-gradle-plugin/4.1.1/224dad42cfe8ff38cfdb42d6e4418da2a231b55cdb1b232e97b0e33c697dbad1/graphql-java-codegen-gradle-plugin-4.1.1.pom'. #16 160.3 > plugins-artifacts.gradle.org: Try again #16 160.3 #16 160.3 * Try: #16 160.3 Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. #16 160.3 #16 160.3 * Get more help at https://help.gradle.org #16 160.3 #16 160.3 Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. #16 160.3 Use '--warning-mode all' to show the individual deprecation warnings. #16 160.3 See https://docs.gradle.org/6.9.2/userguide/command_line_interface.html#sec:command_line_warnings #16 160.3 #16 160.3 BUILD FAILED in 2m 39s
    b
    • 2
    • 1
  • q

    quick-megabyte-61846

    06/27/2022, 1:44 PM
    Hello Iโ€™m trying to ingest lookml using sample projects that looker provides For example, I want to ingest
    welcome_to_looker
    that is provided for demo purposes Iโ€™m getting error:
    Copy code
    1 validation error for LookMLSourceConfig
    base_folder
      file or directory at path "/projects/welcome_to_looker" does not exist (type=value_error.path.not_exists; path=/projects/welcome_to_looker)
    My yaml config:
    Copy code
    base_folder: /projects/welcome_to_looker
    Does that somebody have some tips to specify path correctly?
    o
    • 2
    • 2
  • b

    best-lamp-53937

    06/27/2022, 3:19 PM
    Hi all I am trying to use the Generate documentation for metadata feature Given the
    ./gradelw
    command is relative to the current directory, from which directory should I be issue this command?
    b
    • 2
    • 3
  • g

    gifted-bird-57147

    06/27/2022, 3:26 PM
    I just nuked my local installation and did a new deploy via datahub docker quickstart (v 0.8.39). I can ingest data
    Copy code
    Sink (datahub-rest) report:
    {'records_written': 147,
     'warnings': [],
     'failures': [],
     'downstream_start_time': datetime.datetime(2022, 6, 27, 13, 54, 33, 123014),
     'downstream_end_time': datetime.datetime(2022, 6, 27, 13, 54, 44, 555620),
     'downstream_total_latency_in_seconds': 11.432606,
     'gms_version': 'v0.8.39'}
    but it's not showing on the UI... the compose logs are showing:
    Copy code
    elasticsearch             | {"type": "server", "timestamp": "2022-06-27T15:24:43,298Z", "level": "WARN", "component": "o.e.c.r.a.DiskThresholdMonitor", "cluster.name": "docker-cluster", "node.name": "elasticsearch", "message": "flood stage disk watermark [95%] exceeded on [NT1i9TTQSwW42zsgX84oAw][elasticsearch][/usr/share/elasticsearch/data/nodes/0] free: 2.2gb[2.4%], all indices on this node will be marked read-only", "cluster.uuid": "sQMENsZFR7CcBJCms4Qydg", "node.id": "NT1i9TTQSwW42zsgX84oAw"  }
    is that some issue in the quickstart config?
    b
    • 2
    • 2
  • g

    gray-translator-74287

    06/27/2022, 3:34 PM
    Hello everyone! I am trying to start using analytics page in our locally running Datahub instance (v0.8.39) The first half of the page is pretty empty, but that may be because we haven't been using it that long. However the second half of the analytics page is suspiciously empty, even more than I expected.
    o
    • 2
    • 6
  • a

    agreeable-belgium-70840

    06/27/2022, 3:47 PM
    hello, I am having this issue on datahub v0.8.36. I can see in the main screen that there are 8 glossary terms, but when I click, I can't see any. And the number of glossary terms is not changing when I am adding a new glossary term via the UI. Any ideas?
    o
    b
    b
    • 4
    • 12
  • r

    rich-policeman-92383

    06/27/2022, 8:32 PM
    Hello How do we authenticate when using locust. I have tried specifying the headers but gms logs still show request is missing authorization headers.
    • 1
    • 2
  • b

    breezy-controller-54597

    06/28/2022, 6:30 AM
    I have an error with datahub-mae-consumer in v0.8.39 using helm chart (datahub-0.2.84).
    Copy code
    [main] ERROR o.s.b.d.LoggingFailureAnalysisReporter - 
    
    ***************************
    APPLICATION FAILED TO START
    ***************************
    
    Description:
    
    Parameter 1 of constructor in com.linkedin.metadata.kafka.hook.siblings.SiblingAssociationHook required a bean named โ€˜entityAspectDaoโ€™ that could not be found.
    
    The injection point has the following annotations:
        - @javax.anotation.Nonnull(when =ALWAYS)
    
    Action:
    
    Consider defining a bean named โ€˜entityAspectDaoโ€™ in your configuration.
    I think the following PR is related. https://github.com/datahub-project/datahub/pull/5190
    plus1 3
    b
    s
    +4
    • 7
    • 10
  • b

    bland-orange-13353

    06/28/2022, 9:44 AM
    This message was deleted.
  • m

    microscopic-mechanic-13766

    06/28/2022, 11:28 AM
    Hi, I am trying to deploy Datahub 0.8.39 with ElasticSearch. Although I have specified
    GRAPH_SERVICE_IMPL=elasticsearch
    , I think it doesn't work correctly as in the following error log some lines containing Neo4j can be seen. That is not the only problem, but I have no clue what the rest of the errors' cause. Could someone help me narrow it down??
    o
    a
    • 3
    • 17
  • m

    microscopic-helicopter-87069

    06/28/2022, 6:43 PM
    Hi, is it normal that the upstream in lineage doesn't link to the dataset. Or is it an error in my configuration?
  • b

    busy-machine-97034

    06/29/2022, 1:21 AM
    Hello everyone! I'm getting the following error at the moment to execute an ingestion from Redsfhit:
    acryl.executor.execution.task.TaskError: Failed to resolve secret with name REDSHIFT_HOST_PORT. Aborting recipe execution
    . The datasource was created a month ago and it was working good, but since I upgraded my datahub instance to the 0.3.38 version the data ingestion has been broken. I could't find a solution to this exception, I can't read any secret from that ingestion. Is there any consideration that I missed at the moment of the version upgrade? Hope someone can help me! Thanks!
    l
    c
    • 3
    • 12
  • b

    better-orange-49102

    06/29/2022, 7:19 AM
    im finding that sometimes, upon ingesting a new version of an aspect, the original version of the aspect will disappear from mySQL. But if i were to follow up and emit a slightly different value again, the newly inserted aspect does get incremented correctly. What are some possible areas to check? These aspects are not using ingestion sources, but generated mcps from the python library. I did not set a retention policy yet, and there are no errror messages in the frontend/gms logs. I also didnt specify a version# in the MCP.
    • 1
    • 3
  • s

    silly-ice-4153

    06/29/2022, 9:10 AM
    Hello, I have trouble running the docker images on mac with M1 Chip. It seems that you don"t build the docker images for this architecture. I'm using this for local tests and demo purposes internally. Has someone gotten the docker images to run on this architecture?
    d
    • 2
    • 9
  • h

    high-hospital-85984

    06/29/2022, 10:10 AM
    How do I run the
    ClearSearchServiceStep
    as part of the R`estoreIndices` upgrade command?
    • 1
    • 1
  • p

    polite-actor-701

    06/29/2022, 10:55 AM
    hello. I am trying to use datahub in Korea. I am currently using v0.8.32 installed. When I try to input Korean in urn or description, etc., an error occurs. I tried to install v0.8.36 or v0.8.39, but I couldn't use it with a red error window saying "Oops, an error occurred". I think I need utf8 encoding to input Korean. What should I do?
    o
    d
    • 3
    • 9
  • k

    kind-mechanic-93559

    06/29/2022, 2:06 PM
    Hi Team, My build is failing at ./gradlew metadata ingestioninstallDev can someone help on it. My system is macbook Intel Core i7.
    a
    • 2
    • 3
1...353637...119Latest