Remove disk-and-datanode-size-in-hdfs
article thumbnail

A Reference Architecture for the Cloudera Private Cloud Base Data Platform

Cloudera

Introduction and Rationale. The release of Cloudera Data Platform (CDP) Private Cloud Base edition provides customers with a next generation hybrid cloud architecture. Private Cloud Base Overview. This unified distribution is a scalable and customizable platform where you can securely run many types of workloads. Summary of major changes.

article thumbnail

Automated Deployment of CDP Private Cloud Clusters

Cloudera

At Cloudera, we have long believed that automation is key to delivering secure, ready-to-use, and well-configured platforms. Hence, we were pleased to announce the public release of Ansible-based automation to deploy CDP Private Cloud Base. The most powerful tool we have as developers is automation.” — Scott Hanselman. In this file ( ~/.config/cloudera-deploy/profiles/default

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Cloudera Operational Database Infrastructure Planning Considerations

Cloudera

When operational database is the primary use case in your stack of services, you will need the following: Dedicated storage: Use hard disks that are dedicated to the operational database. You can learn more about Cloudera Operational Database in CDP here: Operational Database in CDP. . Cloudera Operational Database is the primary use case .

OLAP 58
article thumbnail

HBase Performance testing using YCSB

Cloudera

When running any performance benchmarking tool on your cluster, a critical decision is always what data set size should be used for a performance test, and here we demonstrate why it is important to select a “good fit” data set size when running a HBase performance test on your cluster. Operation Database cluster.

Testing 58
article thumbnail

Ozone Write Pipeline V2 with Ratis Streaming

Cloudera

The V2 implementation also avoids unnecessary buffer copying and has a better utilization of the CPUs and the disks in each datanode. When a client writes an object to Ozone, the object is automatically replicated to three datanodes. Ozone is able to scale to billions of objects and hundreds petabytes of data.

article thumbnail

Apache Ozone and Dense Data Nodes

Cloudera

Apache Ozone brings the best of both HDFS and Object Store: Overcomes HDFS limitations. Can support billions of files ( tested up to 10 billion files) in contrast with HDFS which runs into scalability thresholds at 400 million files. Supports 16TB drives unlike HDFS which only supports up to 8 TB drives. Exabyte scale.