You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Currently if we want to develop a new workflow for Nucleus, we have to complete the following steps.
Create and push the docker image to ECR
Create an EFS volume to share data/configuration files between docker containers (optional, only if the use case must share data/configuration between tasks)
Create an ECS task (This can be done through AWS Console or Terraform)
Implement a DAG (workflow) primarily using the Airflow ECS Operator
Upload DAGs to S3 bucket
We have to execute above multiple steps because different PDS docker images have different setups. For example, some of the PDS docker images need various environment variables and others need various volumes to be mounted. As a result, we cannot use a common template.
This problem is there regardless of the workflow tool we use (not specific to Airflow). With any workflow tool, we will have to maintain different setups for different docker images (for example, when we used docker compose, we had different settings for each components in docker compose file).
Implementing and uploading the DAG is easy if the PDS components are already available as ECS tasks. If we plan to automatically generate Terraform scripts, then we have to discuss,
What are the INPUTS that we get from users to generate these scripts?
How should those INPUTS looks like?
Should we maintain a catalog of docker components that they have to select from?
Should we have a UI to select docker components required for a workflow and after making the the selection a way to generate and execute the scripts?
A Workflow Manager Utility can be solution to address above issues.
The text was updated successfully, but these errors were encountered:
💡 Description
Currently if we want to develop a new workflow for Nucleus, we have to complete the following steps.
We have to execute above multiple steps because different PDS docker images have different setups. For example, some of the PDS docker images need various environment variables and others need various volumes to be mounted. As a result, we cannot use a common template.
This problem is there regardless of the workflow tool we use (not specific to Airflow). With any workflow tool, we will have to maintain different setups for different docker images (for example, when we used docker compose, we had different settings for each components in docker compose file).
Implementing and uploading the DAG is easy if the PDS components are already available as ECS tasks. If we plan to automatically generate Terraform scripts, then we have to discuss,
A Workflow Manager Utility can be solution to address above issues.
The text was updated successfully, but these errors were encountered: