Sign In Start Free Trial
Account

Add to playlist

Create a Playlist

Modal Close icon
You need to login to use this feature.
  • HBase High Performance Cookbook
  • Toc
  • feedback
HBase High Performance Cookbook

HBase High Performance Cookbook

By : Ruchir Choudhry
2.5 (2)
close
HBase High Performance Cookbook

HBase High Performance Cookbook

2.5 (2)
By: Ruchir Choudhry

Overview of this book

Apache HBase is a non-relational NoSQL database management system that runs on top of HDFS. It is an open source, disturbed, versioned, column-oriented store and is written in Java to provide random real-time access to big Data. We’ll start off by ensuring you have a solid understanding the basics of HBase, followed by giving you a thorough explanation of architecting a HBase cluster as per our project specifications. Next, we will explore the scalable structure of tables and we will be able to communicate with the HBase client. After this, we’ll show you the intricacies of MapReduce and the art of performance tuning with HBase. Following this, we’ll explain the concepts pertaining to scaling with HBase. Finally, you will get an understanding of how to integrate HBase with other tools such as ElasticSearch. By the end of this book, you will have learned enough to exploit HBase for boost system performance.
Table of Contents (13 chapters)
close
7
7. Large-Scale MapReduce
12
Index

Bulk utilities

The process for loading data using Bulk utilities is very similar:

  1. Extracting data from the source.
  2. Transforming the data into HFiles.
  3. Loading the files into HBase by guiding the region servers as to where to find them.

Getting ready...

The following points have to be remembered when using Bulk utilities:

  • HBase/Hadoop cluster with MapReduce/Yarn should be running. You can run jps to check it.
  • Access rights (user/group) are needed to execute the program.
  • Table schema needs to be designed to the input structure.
  • Split points need to be taken into consideration.
  • The entire stack (compaction, split, block size, max file size, flush size, version compression, mem store size, block cache, garbage collections nproc, and so on) needs to be fine-tuned to make the best of it.

The WAL are not written; thus, data lost during the failure may not be recovered as there is no replication performed by reading the WAL.

How to do it…

There are multiple ways to do this work, such as writing your own...

Unlock full access

Continue reading for free

A Packt free trial gives you instant online access to our library of over 7000 practical eBooks and videos, constantly updated with the latest in tech
bookmark search playlist download font-size

Change the font size

margin-width

Change margin width

day-mode

Change background colour

Close icon Search
Country selected

Close icon Your notes and bookmarks

Delete Bookmark

Modal Close icon
Are you sure you want to delete it?
Cancel
Yes, Delete