article thumbnail

Monitor data pipelines in a serverless data lake

AWS Big Data

The combination of a data lake in a serverless paradigm brings significant cost and performance benefits. By monitoring application logs, you can gain insights into job execution, troubleshoot issues promptly to ensure the overall health and reliability of data pipelines.

article thumbnail

Measure performance of AWS Glue Data Quality for ETL pipelines

AWS Big Data

In recent years, data lakes have become a mainstream architecture, and data quality validation is a critical factor to improve the reusability and consistency of the data. In this post, we provide benchmark results of running increasingly complex data quality rulesets over a predefined test dataset.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

How HR&A uses Amazon Redshift spatial analytics on Amazon Redshift Serverless to measure digital equity in states across the US

AWS Big Data

This dynamic tool, powered by AWS and CARTO, provided robust visualizations of which regions and populations were interacting with our survey, enabling us to zoom in quickly and address gaps in coverage. Figure 1: Workflow illustrating data ingesting, transformation, and visualization using Redshift and CARTO.

article thumbnail

Visualize data quality scores and metrics generated by AWS Glue Data Quality

AWS Big Data

AWS Glue Data Quality allows you to measure and monitor the quality of data in your data repositories. It’s important for business users to be able to see quality scores and metrics to make confident business decisions and debug data quality issues. In the following sections, we discuss these steps in more detail.

article thumbnail

Why the Data Journey Manifesto?

DataKitchen

We had been talking about “Agile Analytic Operations,” “DevOps for Data Teams,” and “Lean Manufacturing For Data,” but the concept was hard to get across and communicate. I spent much time de-categorizing DataOps: we are not discussing ETL, Data Lake, or Data Science. The post Why the Data Journey Manifesto?

Testing 130
article thumbnail

DataOps Observability: Taming the Chaos (Part 3)

DataKitchen

An effective DataOps observability solution requires supporting infrastructure for the journeys to observe and report what’s happening across your data estate. Logs and storage for problem diagnosis and visualization of historical trends. Data and tool tests. And she’ll know when newer data will arrive.

Testing 130
article thumbnail

Set up advanced rules to validate quality of multiple datasets with AWS Glue Data Quality

AWS Big Data

Poor-quality data can lead to incorrect insights, bad decisions, and lost opportunities. AWS Glue Data Quality measures and monitors the quality of your dataset. It supports both data quality at rest and data quality in AWS Glue extract, transform, and load (ETL) pipelines.