article thumbnail

Orca Security’s journey to a petabyte-scale data lake with Apache Iceberg and AWS Analytics

AWS Big Data

With data becoming the driving force behind many industries today, having a modern data architecture is pivotal for organizations to be successful. This ensures that the data is suitable for training purposes. These robust capabilities ensure that data within the data lake remains accurate, consistent, and reliable.

article thumbnail

Create a modern data platform using the Data Build Tool (dbt) in the AWS Cloud

AWS Big Data

In this post, we delve into a case study for a retail use case, exploring how the Data Build Tool (dbt) was used effectively within an AWS environment to build a high-performing, efficient, and modern data platform. It does this by helping teams handle the T in ETL (extract, transform, and load) processes.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Deep dive into the AWS ProServe Hadoop Migration Delivery Kit TCO tool

AWS Big Data

Let’s look at some key metrics. After analyzing YARN logs by various metrics, you’re ready to design future EMR architectures. He also understands how to apply technologies to solve big data problems and build a well-designed data architecture. George Zhao is a Senior Data Architect at AWS ProServe.

article thumbnail

BMW Cloud Efficiency Analytics powered by Amazon QuickSight and Amazon Athena

AWS Big Data

The difference lies in when and where data transformation takes place. In ETL, data is transformed before it’s loaded into the data warehouse. In ELT, raw data is loaded into the data warehouse first, then it’s transformed directly within the warehouse.

article thumbnail

How smava makes loans transparent and affordable using Amazon Redshift Serverless

AWS Big Data

Overview of solution As a data-driven company, smava relies on the AWS Cloud to power their analytics use cases. smava ingests data from various external and internal data sources into a landing stage on the data lake based on Amazon Simple Storage Service (Amazon S3).

article thumbnail

Introducing the AWS ProServe Hadoop Migration Delivery Kit TCO tool

AWS Big Data

The results of the log analyzer reveal Hadoop workload insights with various views and metrics of the Hadoop applications shown in Amazon QuickSight dashboards, which leads to the design of a future EMR cluster. He also understands how to apply technologies to solve big data problems and build a well-designed data architecture.

article thumbnail

Birst automates the creation of data warehouses in Snowflake

Birst BI

Customers such as Crossmark , DJO Global and others use Birst with Snowflake to deliver the ultimate modern data architecture. Data never leaves Snowflake with Birst’s ability to support the reporting and self-service needs of both centralized IT and decentralized LOB teams.