AWS Tutorials - Data Quality Check using AWS Glue DataBrew
HTML-код
- Опубликовано: 20 ноя 2021
- The code link - github.com/aws-dojo/analytics...
Maintaining data quality is very important for the data platform. Bad data can break ETL jobs. It can crash dashboards and reports. It can hit accuracy of the machine learning models due to bias and error. AWS Glue DataBrew Data Profile jobs can be used for data quality checks. One can define data quality rules and validate data against it. Learning how to use Data Quality Rules in AWS Glue DataBrew to validate data quality.
Thanks, very comprehensive overview of the quality checking in DataBrew.
That was extremely helpful, thank you!
Very nicely explained..
Nice explanation and details
Very impressive, I have been looking at data validation frameworks and think this would be great fit. The 2 open source libraries I checked are:
Thank you for the tutorial which can have understanding on the overall about the DQ part. Whether having possible to view the detail records which is succeeded or fail?
I'm looking for the most code-light (a short Python Lambda function is ok and assumed) way to set up a process so when a CSV file is dropped into my S3 bucket/incoming folder, the file will automatically be validated using a DQ Ruleset I would manually build earlier in console. For any given Lambda call (I assume triggered by a file dropped into our S3 bucket) If possible, I'd like the Lambda to instruct the DQ Ruleset to run but not wait for it to finish (Step function?). Wanting to output a log file of which rows/columns failed to my S3 bucket/reports folder (Using some kind of trigger that fires from a DQ Ruleset finishing execution?). Again, it is important that the process be fully automated because hundreds of files per day with hundreds of thousands of rows will be dropped into our S3 bucket/incoming folder every day via a different automated process. End goal is merely to let client know if their file does not fit rules. No need to save or clean data. I realize I may be asking a lot, so please feel free to only share the best high level path of which AWS services to use in which order. Thank you!
This was very nicely explained! Thank yo so much :)
This is perfect. We have thousands of datasets where we need to perform DQ checks and send reports. Is it possible to automate or create the rules programmatically instead of using the console? Something like create rules in a yaml/csv file?
Great.. Do have any plans to make a video on aws glue and apache hudi integration?
Thanks for the clear explanation!
where you have placed this code and how it is connected with this data brew profile job
Its nice explaination any training you will give I am looking to training pls help me ...
Please made video on pydeequ with Glue -> without using EMR
Can you pls give training for aws glue we are 5 members looking for training