jan_soubusta
02/01/2024, 11:05 PMAndy Carter
02/12/2024, 3:51 PMdbt_packages
dir was not getting rebuilt and just picking up my local copy.
How can I run a dbt deps
command in my dockerfile if dbt is installed as part of my dagster utility env? Here's the relevant section of my meltano.yml
utilities:
- name: dagster
variant: quantile-development
pip_url: dagster-ext dagster-postgres dagster-dbt dbt-postgres dagster-azure pendulum==2.1.2 dagster_msteams
settings:
- name: dagster_home
env: DAGSTER_HOME
value: $MELTANO_PROJECT_ROOT/orchestrate/dagster
commands:
dev:
args: dev -f $REPOSITORY_DIR/repository.py --dagit-host 0.0.0.0 -d $REPOSITORY_DIR
executable: dagster_invoker
choudary kukkapalli
02/12/2024, 8:14 PMmeltano==3.3.1
--> this is what in requirements file
FROM python:3.11-slim
LABEL maintainer="WFS Corp"
# Set working directory
WORKDIR /opt
# Install OS dependencies
RUN apt-get update && \
apt-get install -y build-essential freetds-bin freetds-dev git libkrb5-dev libssl-dev tdsodbc unixodbc unixodbc-dev && \
rm -rf /var/cache/apt/archives /var/lib/apt/lists/*
# ssl cert
COPY certs/zscaler.pem /usr/local/share/ca-certificates/zscaler.crt
RUN update-ca-certificates --fresh
#ENV HTTP_PROXY=
#ENV HTTPS_PROXY=
ENV REQUESTS_CA_BUNDLE=/usr/local/share/ca-certificates/zscaler.crt
ENV CURL_CA_BUNDLE=/usr/local/share/ca-certificates/zscaler.crt
# Make sure we are using latest pip
RUN pip install --upgrade pip wheel
# Copy requirements.txt
COPY ../requirements.txt requirements.txt
# Install dependencies
RUN pip install -r requirements.txt
Ian OLeary
03/12/2024, 6:40 PMyao_zhang
03/19/2024, 11:34 PMmeltano/meltano:v3.3-python3.10
and have a plugin that is dependent on python3.9
while the rest are compatible with 3.10
. I want to structure the dockerfile so that it supports multiple python versions and do meltano install
of the plugins during different build stages with different python versions. Can someone show me how the dockerfile should be structured or provide an example? https://docs.meltano.com/reference/settings/#pythonChris Goodell
03/21/2024, 1:26 PMtap-linkedin-ads
extractor (meltanolabs implementation) to obtain LinkedIn Ads data, and I am a little curious about how the token refresh is intended to be managed?
When you obtain an OAuth 2.0 access token for Linkedin, you also are provided a refresh token in order to make a secondary call to refresh your access token. (The access token has an expiry of 60 days, and the refresh token has an expiry of 365 days.)
I didn't see anything in the client.py where the refresh token is actively managed, so the access token would expire in 60 days by default, unless I am missing something? Is the tap able to manage this somehow?
I also looked at the other versions for this same `tap-linkedin-ads`; the Stitch Data implementation seems to have a provision for that functionality in client.py, in two functions named refresh_access_token
and fetch_and_set_access_token
, but I am unsure how the new access token would be stored, as it would refresh the token and receive the new one, but it is self contained within the k8s pod and won't persist.
https://hub.meltano.com/extractors/tap-linkedin-ads
https://github.com/MeltanoLabs/tap-linkedin-ads
https://github.com/singer-io/tap-linkedin-adsWilli Langlitz
03/22/2024, 11:44 AMWilli Langlitz
03/22/2024, 11:45 AMWilli Langlitz
03/22/2024, 11:45 AMSiva Achyuth
04/02/2024, 10:57 AMFayaz Ahmed
05/06/2024, 6:50 AMjoshua_janicas
05/09/2024, 7:16 PMCMD
or ENTRYPOINT
, nothing seems to happen and Docker composes without ever starting Dagster. Looking for thoughts as to what I could be doing wrong here.Andy Carter
05/16/2024, 11:01 AMSiddu Hussain
05/29/2024, 5:12 AMsreepriya m
05/31/2024, 3:01 PMsreepriya m
05/31/2024, 3:06 PMsreepriya m
05/31/2024, 3:09 PMAhmed Hamid
06/02/2024, 10:10 AMSiddu Hussain
06/03/2024, 7:53 PMsreepriya m
06/12/2024, 6:43 PMsreepriya m
06/21/2024, 11:21 AMJens Christian Hillerup
07/14/2024, 5:49 PMtap-postgres
and target-postgres
up and running pretty quickly and I'm ready to try deploying it to Heroku, but I'm wondering about the .meltano
directory: besides the meltano.db SQLite database, does it contain anything that must be persisted?
This docs page lists what's in the directory:
β’ I can live without the log files for prod (or potentially find a way to get the logs themselves extracted and loaded somehow)
β’ I suppose the `venv`s of the needed Python packages could be created at docker build
-time?
I know Meltano supports pluggable system databases, and I'm planning on just using letting Meltano have a schema in my BI database for that. Other than that, what else do I need to know for a stateless Docker deployment (on Heroku, in my case)?Matthew Hooson
08/10/2024, 10:28 AMhaleemur_ali
09/20/2024, 1:52 PMNghia Nguyen Truong Tri
11/07/2024, 5:21 PMNghia Nguyen Truong Tri
11/07/2024, 5:23 PMCreating system database...Need help fixing this problem? Visit <http://melta.no/> for troubleshooting steps, or to
join our friendly Slack community.
Failed to initialize database: (sqlite3.OperationalError) disk I/O error
[SQL: PRAGMA journal_mode=WAL]
(Background on this error at: <https://sqlalche.me/e/20/e3q8>)
Abednego Santoso
11/26/2024, 2:35 PMmeltano --version
, it said meltano: command not found
. Can you please help me?Jacob Ukokobili
01/21/2025, 2:01 PMversion: 1
default_environment: dev
project_id: 751cca76-711b-46ec-8e5c-26afb7f94623
environments:
- name: dev
- name: staging
- name: prod
plugins:
extractors:
- name: tap-mysql
variant: transferwise
pip_url: git+<https://github.com/transferwise/pipelinewise.git#subdirectory=singer-connectors/tap-mysql>
config:
database: ${TAP_MYSQL_DATABASE}
user: ${TAP_MYSQL_USER}
port: ${TAP_MYSQL_PORT}
host: ${TAP_MYSQL_HOST}
select:
'*.*': true # Select all tables by default
metadata:
'*.*':
replication-method: INCREMENTAL # Use INCREMENTAL replication for all tables
replication_key: update_time # Replace with your timestamp column
key_properties:
- id # Replace with your primary key column
'*.*_audit':
selected: false # Exclude tables ending with "_audit"
loaders:
- name: target-bigquery
variant: z3z1ma
pip_url: git+<https://github.com/z3z1ma/target-bigquery.git>
config:
dataset: ${TARGET_BIGQUERY_DATASET}
location: ${TARGET_BIGQUERY_LOCATION}
project: ${TARGET_BIGQUERY_PROJECT}
credentials_json: ${TARGET_BIGQUERY_CREDENTIALS_JSON}
Victor Castro
01/23/2025, 12:46 AMMario
02/06/2025, 12:30 AMtap-braze
plugin which relies on Airbyte. Recently (as of this PR it seems), airbyte stopped making their pypi package available (based on their registry), forcing meltano to use the docker image.
I currently use our existing Airflow to orchestrate meltano jobs. A KubernetesPodOperator
using the following arguments
arguments=["run", "tap-braze", "target-postgres"],
runs the image that we built which contains our project. Up until about a week ago, when said PR was merged, this was working just fine (I assume because it was using the pypi package with no problem). Now the tap-braze
plugin fails with launching the airbyte docker image since we're running our Airflow cluster on EKS.
The Airbyte wrapper assumes usage of docker
even though OCI_RUNTIME
is being overridden. I'm running on EKS 1.31 which no longer uses docker as a runtime but rather containerd
. I've worked around this by installing nerdctl
into my project docker image (as it should be a drop-in replacement for docker cli) and `ln`'ing nerdctl
to docker
.
I've also gone ahead and mounted the following host directories to get nerdctl
to at least be able to pull the airbyte image.
β’ /tmp
β’ /run/containerd
β’ /var/lib/containerd
β’ /var/lib/nerdctl
My problem is that I still get the following error message and I'm kinda lost lol, my next guess is that the host running the airbyte image needs to have nerdctl
installed but I'm hoping someone else has come across something similar and solved this in a different way.
time="2025-02-06T000356Z" level=fatal msg="failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: fork/exec /usr/bin/nerdctl: no such file or directory: unknown"