Skip to content

IndustrialDataops/argoflow

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Pipeline API

Argo Workflow automation is driven by yaml templates that are desined using the ADSL (Argo Domain Specific Language). More examples for the same can be found here

Pipeline API makes use of the Argo models defined in the Argo Python client repository

Why Pipeline API

  • Hide infrastructure/underlying image details from the end user
  • Embedding python code into other existing automation scripts

All the low level details regarding the image/container details are stored in yaml file

Configuration file :

Config file contains mainly two components

  • Containers : contains the pre built container images
  • Resources : can contain any custom resource definition that needs to be implemented

Below is the sample yaml which contains jobprofilerclient to run a sample data profiler using pydeequ and resource template sparkk8sScala is to run a spark job using Spark Operator

ConfigFile

Usage

A quick example to show how to complex DAG can be created

tasks

Once tasks has been added , dag can be compiled and also we can view the dependencies once the dag is compiled , submit the dag to argo workflow

workflow

This code results in a DAG which looks like this when submitted to ArgoWorkflow

dag