Measure and Monitor Data Quality of your Datasets in AWS Glue Data Catalog | Amazon Web Services

preview_player
Показать описание
AWS Glue Data Quality is a feature of AWS Glue that measures and monitors the data quality of data repositories. In this video, we provide an overview of what AWS Glue Data Quality is, key capabilities, and walk you through how you could use it to manage data quality for your data assets cataloged in AWS Glue Data Catalog. You will gain a clear understanding of how to use this capability after you review this video.

Subscribe:

Do you have technical AWS questions?

ABOUT AWS
Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform, offering over 200 fully featured services from data centers globally. Millions of customers — including the fastest-growing startups, largest enterprises, and leading government agencies — are using AWS to lower costs, become more agile, and innovate faster.

#AWS #AmazonWebServices #CloudComputing #DataQuality #DataGovernance #AWSGlue #AWSGlueDataQuality #ETL #DataCatalog #GlueStudio #DataIntegration
Рекомендации по теме
Комментарии
Автор

Question: I have a table1 (stored in AWS Apache Iceberg) with precalculated data. Then I run a Glue Job that uses this table1 in some SQL query (join with other tables in query) and copy data produced by this sql query into Aurora DB. I want to use Dataguallity to verify ( match several columns by values, to make sure that value from the column of source table1 is equal to the value from the column in the destination Aurora table) if I copied the same data as it was in the source table. How I can do that using AWS Glue Data Quality?

petrodyak