Skip to content

A Dagster-based data transformation pipeline, used to fetch, convert, validate & publish datasets.

License

Notifications You must be signed in to change notification settings

mobidata-bw/ipl-dagster-pipeline

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

IPL Data Pipeline

This repo represents the data transformation pipeline of the MobiData-BW Integrationsplatform (IPL).

It uses Dagster to retrieve and transform several datasources. The results are either published as datasets, or written into databases and then being served (e.g. as WMS/WFS or REST service) by other IPL services.

Usage

Note: This repo to is designed to be run as a part of the entire IPL platform, as defined in the ipl-orchestration repo. But you can also run it in a standalone fashion.

Prerequisites

If you intend to run this project locally via dagster dev, you need to have python 3.10 or 3.11 installed. Python 3.12 is not yet supported by dagster.

To install all required libraries, use

pip install -r requirements.txt

Note: requirements.txt imports requirements-pipeline.txt and requirements-dagster.txt, which include the dependencies for the different dagster services. pipeline.Dockerfile and dagster.Dockerfile just import these respective requirements.

In addition, you need a postgres database into which the datasets are loaded. This database can be started via

docker compose -f docker-compose.dev.yml up

Running

To start this dagster project in interactive develepment mode, you should use a DAGSTER_HOME other than this project directory, as a) the dagster.yml defines a postgres storage for dagster run information and is usually intended for prod use, and b) a number of files is generated in the temporary dagster directories which would impact your IDE responsiveness if it's indexing new files continuously.

$  DAGSTER_HOME=/tmp/DAGSTER_HOME dagster dev

or via docker compose, which is the way it is itended to be deployed with:

$ docker compose up --build

Note that the config differs in that for docker-compose, workspace.docker.yaml and dagster.docker.yaml will be used, which configure a postgres db as dagster storage, whild dagster dev will use sqlite and temporary folders.

About

A Dagster-based data transformation pipeline, used to fetch, convert, validate & publish datasets.

Resources

License

Stars

Watchers

Forks

Packages

 
 
 

Contributors 4

  •  
  •  
  •  
  •