-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Validate and Load all PDS4 MESSENGER data products with Nucleus #54
Comments
@ramesh-maddegoda focuses on MSGRMDS_4001 and MESSDEM_1001 that need to be loaded in the registry first so that they can be used in ticket NASA-PDS/search-api-notebook#24 |
Blocked because AWS Airflow is unavailable on NGAP |
Unblocked since Ramesh work on MCP. He is now testing the ECS task called by the nucleus workflow. |
Status: @ramesh-maddegoda working on improving Terraform deployments |
@ramesh-maddegoda is deploying everything needed on MCP, from scratch. |
@ramesh-maddegoda will test nucleus to validate its robustness with a bigger dataset. |
Some of the files in the s3://asc-pds-messenger failed to copy to the PDS Nucleus staging bucket with a permission issue.
|
A new parameter enable to copy all the metadata . |
…e verified files DataSync report instead of the transferred files report Refer to task #54
…ion lambda code to make sure both product table product_data_file_mapping table are updated in a consistent way (make sure both tables are updated). Refer to task #54
…ion lambda code to make sure both product table product_data_file_mapping table are updated in a consistent way (make sure both tables are updated). Refer to task #54
@ramesh-maddegoda identified a bug while doing that test. The lambda reading the data sync report is now taking more that 15 minutes. Now there will be a single lambda call per report. |
The upgrade worked on a small dataset and @ramesh-maddegoda is now testing on the messenger dataset. |
…ing to process a large amount of files in one lambda Refer to task #54
…ing to process a large amount of files in one lambda Refer to task #54
…ing to process a large amount of files in one lambda Refer to task #54
Now Ramesh is loading data to the registry on JPL AWS. Last step for this task. |
20,000 processed! 8 directories ran! Found 2 errors:
|
|
@ramesh-maddegoda is experimenting with SQS to send to new records to the mysql database and avoid the time out he was experiencing with direct insertion. |
SQS now mostly works, but another lambda had a time out. |
We now integrate the copy from S3 to EFS as a nucleus step in the DAGs. We give up DataSync which comes with risks of overlapping copies and complication to remove files from EFS. @ramesh-maddegoda will also write a note in a wiki for a future design where we don't need to use EFS at all. |
This work has been paused as we focus on Catalina Sky Survey. Will move to B15.0 release plan to complete work. |
📆 05/2024 status: Delayed several sprints due to delays in #93. This is an operations activity. No impact on build. |
📆 06/2024 status: Delayed several sprints due to delays in #93. This is an operations activity. No impact on build. |
📆 07/2024 status: Delayed several sprints due to delays in #93. This is an operations activity. No impact on build. |
📆 08/2024 status: Delayed several sprints due to delays in #93. This is an operations activity. No impact on build. Will most likely be deferred to B15.1 |
📆 09/2024 status: deferred to B15.1 due to delays in #93. This is an operations activity. No impact on build. |
💡 Description
The text was updated successfully, but these errors were encountered: