Git Product home page Git Product logo

gooddata-data-pipeline's Introduction

GoodData Data Pipeline

The demo inside this repository demonstrates e2e data pipeline following best software engineering practices. It realizes the following steps:

  • crawls data from sources with (Meltano)
  • loads data into a warehouse with (Meltano)
  • transforms data in the warehouse in a multi-dimensional model ready for analytics with (dbt)
  • generates semantic model from physical model (GoodData model from dbt models)
  • deploys analytics model (metrics, insights, dashboards) from locally stored layout files
  • deploys UI data apps coupled with the data pipeline. Read more details in apps folder

Currently, these data warehouse engines are supported:

  • PostgreSQL
  • Vertica
  • Snowflake
  • MotherDuck (partially, TODO)

Delivery into dev/staging/prod environments is orchestrated by GitHub or Gitlab (except the data apps).

Data apps are delivered by render.com service after merge to main branch (staging), and after merge to prod branch(production). Also, a preview URL is generated by before merge as a comment in the MR in Gitlab. Note: Streamlit cloud does not support Gitlab.

Generally, you can change the whole data pipeline and UI apps by a single commit, and deliver everything consistently.

Architecture picture

Demo architecture

If you need help

This README is just a brief how-to, it does not contain all details. If you need help, do not hesitate to ask in our Slack community.

Authors

Related articles

The following articles are based on this project:

Getting Started

I recommend to begin on your localhost, starting the whole ecosystem using docker-compose.yaml file. It utilizes the GoodData Container Edition available in DockerHub. Optionally, you can also start Vertica Community Edition.

WARNING: Since version 3.0.0, the GoodData.CN Community Edition will be renamed to GoodData.CN Container Edition, and will require a license key for its run. If you are a GoodData customer, you can use the same license key as for the production deployment. If not, please, contact us in our Slack Community.

Personal note: I was fighting with the above decision for a long time, but finally, I lost. I encourage you to ask for a return of a free Community Edition in the Slack community. The more people ask, the more chances we have to get it back.

# Build custom images based on Meltano, dbt and GoodData artefacts
docker-compose build
# Start GoodData, and Minio(AWS S3 Meltano state backend)
docker-compose up -d gooddata-cn-ce minio minio-bootstrap
# Wait 1-2 minutes to services successfully start
# Optionally, start Vertica
docker-compose up -d vertica

# Allow https://localhost:8443 in CORS
# This enables testing of locally started UI apps based on UI.SDK (examples in /apps folder) 
docker-compose up bootstrap_origins

# Extract/load pipeline based on Meltano
# Github token for authenticating with Github REST API 
export TAP_GITHUB_AUTH_TOKEN="<my github token>"
# If you use Vertica
export VERTICA_PASS"<your Vertica password>" 
# If you use MotherDuck
export MOTHERDUCK_TOKEN="<your MotherDuck token>" 
docker-compose up extract_load_github
docker-compose up extract_load_faa
# Uncomment once the tap is fixed
# docker-compose up extract_load_exchangeratehost
docker-compose up extract_load_data_science
docker-compose up extract_load_ecommerce_demo

# Transform model to be ready for analytics, with dbt
# Also, GoodData models are generated from dbt models and pushed to GoodData  
docker-compose up transform  

# Deliver analytics artefacts(metrics, visualizations, dashboards, ...) into GoodData
docker-compose up analytics

Once everything finishes successfully, you can go to http://localhost:3000, log in with [email protected]/demo123, and start consuming the result of the data pipeline in the form of dashboards.

There is also an alternative consumption experience - you can start data apps stored in apps folder. check corresponding README files in sub-folders each representing a particular app.

Getting started - cloud

Move to Gitlab, fork this repository and run the pipeline against your environments:

  • Create a public GoodData instance
    • Go to GoodData trial page, enter your e-mail, and in few tens of seconds you get your own GoodData instance running in our cloud, managed by us.
  • Create a public PostgreSQL or Snowflake instance
    • Create required databases (for dev/staging/prod).

You have to set the following (sensitive) environment variables in the Gitlab(section Settings/CICD):

  • TAP_GITHUB_AUTH_TOKEN
  • GoodData
    • Single endpoint, host+port
      • GOODDATA_HOST - host name pointing to the GoodData instance
      • GOODDATA_TOKEN - admin token to authenticate against the GoodData instance
    • Multiple endpoints
      • Create file ~/.gooddata/profiles.yaml
      • Check Example
  • MELTANO_STATE_AWS_ACCESS_KEY_ID/MELTANO_STATE_AWS_SECRET_ACCESS_KEY - Meltano stores its state to AWS S3, and needs these credentials
  • GITLAB_TOKEN - to allow Gitlab jobs to send messages to merge requests
  • GITHUB_TOKEN - to allow GitHub jobs to send messages to pull requests

The rest of environment variables (Github repos to be crawled, DB endpoints, ...) can be configured in .gitlab-ci.yml(section variables). The same variables must be set in GitHub variables UI if you want to use GitHub.

Developer guide

Bootstrap developer environment:

# Creates virtualenv and installs all dependencies
make dev

# Activate virtualenv for extract_load part or for transform/analytics parts
source .venv_el/bin/activate
# Activate virtualenv for transform and analytics parts
source .venv_t/bin/activate
# You should see e.g. a `(.venv_el)` appear at the beginning of your terminal prompt indicating that you are working inside the `virtualenv`.

# Deactivate virtual env once you are done
deactivate

Set environment variables

See .env.local and similar examples. Add sensitive variables to the corresponding .env.custom.<local|dev|staging|prod> file.

source .env.local
# For vertica
source .env.local vertica
# For MotherDuck
source .env.local motherduck
# For Vertica in the cloud
source .env.staging vertica

Extract and Load

Meltano tool is used. Configuration file meltano.yml declares everything related.

How to run:

make extract_load
# Full refresh
FR="--full-refresh" make extract_load

Several jobs are running for each combination of source(tap) and target. The output of this stage are multiple DB schemas. Schema names are declared in .env.local.

Jobs are running incrementally. They store their states into AWS S3 (or local Minio). You can use --full-refresh flag to enforce full refresh of the whole model.

Data transformation

The folder data_pipeline/models contains dbt models to transform data from all schemas produced by the previous step to cicd_output_stage schema that is used for analytics of data. You can use --full-refresh flag to enforce full refresh of the whole model.

How to run:

make transform
# Full refresh
FR="--full-refresh" make transform
# Dry run
FR="--dry-run" make transform

Generate GoodData semantic model from dbt models

Plugin gooddata-dbt provides generators of GoodData semantic model objects from dbt models. In particular, it allows you to generate so called GoodData LDM(Logical Data Model, mapped to tables/columns), and metrics. It is based on GoodData Python SDK.

How to run:

make deploy_models
# There are related make target executing atomic steps

# Dry run
FR="--dry-run" make deploy_models

For Vertica, we have to customize VERTICA_HOST variable, because Vertica is running inside the docker-compose network. When you run e.g. Meltano from localhost, you connect to localhost. When GoodData is connecting to Vertica inside the docker network, it must connect to docker hostname of Vertica, which is vertica.

VERTICA_HOST="vertica" gooddata-dbt deploy_models

Constraints

We wanted to define database constraints. Why? GoodData can join automatically between tables if you have tables that have foreign keys. This is a huge benefit that saves time, and also you can avoid mistakes thanks to that.

We used package Snowflake-Labs/dbt_constraints for defining constrains.

Another option is to declare these semantic properties into GoodData-specific meta sections in dbt models (it is utilized in this demo).

Schema names

The dbt autogenerates the schema name but you can easily change it by custom macro - see our approach how we dealt with schema names.

Generate all columns into schema.yml files

This can help you to bootstrap schema.yaml files programmatically. Then, you can extend them by additional properties.

Example:

dbt --profiles-dir profile run-operation generate_source \
  --args "{\"schema_name\": \"$INPUT_SCHEMA_GITHUB\", \"generate_columns\": true, \"include_descriptions\": true}"

Analytics

Folder gooddata-dbt contains a PoC of dbt plugin providing tools for loading/storing GoodData analytics model (metrics, insights, dashboards). It is based on GoodData Python SDK.

Load analytics model to GoodData

Analytics model is stored in gooddata_layouts folder.

The following command reads the layout, and loads it into the GooData instance:

make deploy_analytics
# Dry run
FR="--dry-run" make deploy_analytics

It not only loads the stored layout, but it also reads metrics from dbt models and loads them too.

Store analytics model

Anytime you can fetch analytics model from the GoodData instance and store it to gooddata_layouts folder. It makes sense to do some operations by editing stored layout files, but others in GoodData UI applications. For instance, it is more convenient to build more complex GoodData MAQL metrics in the Metric Editor UI application. Then, to persist such metrics to gooddata_layouts folder, run the following command:

GOODDATA_PROFILES="local" make store_analytics

WARNING: if you utilize ~/.gooddata/profiles.yaml, and you would fill GOODDATA_PROFILES variable with multiple values, it will store analytics from all related GoodData instances into the same folder. This is not what you usually need. That is why I override GOODDATA_PROFILES variable with a single value.

Invalidate analytics cache

Anytime you update data, e.g. by running dbt run command, you have to invalidate GoodData caches to see the new data there. The following command invalidates these caches:

make invalidate_caches

Test analytics

It is possible to test if all insights (visualizations) are possible to execute - it means that you know if insights (visualizations) render correctly.

Use the following command:

make test_insights

Geo charts

It is necessary to upload Mapbox token to GoodData deployments other than GoodData.cloud (SaaS). Use helper script for that:

cd data_pipeline
export MAPBOX_ACCESS_TOKEN="<put your token here>"
python upload_mapbox_token.py -gw faa_development
# If localization is used
python upload_mapbox_token.py -gw faa_development_fr 
python upload_mapbox_token.py -gw faa_development_chinese__simplified_ 

Applications

Applications are stored in apps folder. They are not delivered by the Gitlab pipeline, but by render.com service watching this repo.

VR demo

README

Streamlit demo

README


If you want to learn more about the stack described in the document, do not hesitate to contact us.

If you find a bug, please create a merge request, or create an issue.

gooddata-data-pipeline's People

Contributors

jaceksan avatar hkad98 avatar patrikbraborec avatar no23reason avatar mara3l avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.