Case 1 of HBase custom HBase MapReduce

1. Demand scenario Migrate part of the data in the ys table in HBase to the ys? Mr table through Mapreduce 2. Coding 1) build ReadysMapreduce class to read data in ys table package cn.ysjh; import java.io.IOException; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.CellUtil; import org.apache.hadoop.hb ...

Posted by jacobsdad on Wed, 11 Dec 2019 05:50:12 +0100

Typical scenes of hbase

1. hbase integration Mapreduce    in the offline task scenario, MapReduce accesses HBASE data to speed up analysis and expand analysis capabilities.Read data from hbase (result) public class ReadHBaseDataMR { private static final String ZK_KEY = "hbase.zookeeper.quorum"; private static final String ZK_VALUE = "hadoop01:2181,h ...

Posted by diagnostix on Wed, 04 Dec 2019 11:55:27 +0100

HBase custom MapReduce

Transfer of HBase table data In the Hadoop phase, the MR task we wrote has two classes: Mapper and Reducer. In HBase, we need to inherit two classes: TableMapper and TableReducer. Objective: to migrate part of the data in the fruit table to the fruit_mr table through MR Step 1. Build the ReadFruitMapper class to read the data in the fruit ta ...

Posted by brooky on Sun, 03 Nov 2019 19:16:43 +0100

Integration of SparkPython and Hbase for Real-time Computing of Large Data Series

1. Preparations (the tool library used will be placed at the end for download) 1.1. Install thrift   cmd>pip install thrift I use Anaconda3. The downloaded packages will be stored in the / Lib/site-packages / directory. If you don't use Anaconda3, you can put the following two folders directly ...

Posted by JacobYaYa on Tue, 01 Oct 2019 22:21:56 +0200

Nutch2.1+Hbase+Solr to quickly build a crawler and search engine (fast, basically within 2 hours)

Note: This method is for quick experience or small data volume, not suitable for large data volume production environment. Environmental preparation: Centos7 Nutch2.2.1 JAVA1.8 ant1.9.14 HBase 0.90.4 (stand-alone version) solr7.7 Relevant download address: Links: https://pan.baidu.com/s/1Tut2CcKoJ9-G-HBq8zexMQ Extraction code: v75v Start inst ...

Posted by MAXIEDECIMAL on Thu, 26 Sep 2019 07:49:34 +0200

HBase filter

Filter There are many types of filters, but they can be divided into two categories - Comparative filters and special filters. The function of the filter is to judge whether the data satisfy the condition at the server side, and then return the data satisfying the condition to the client side only. ...

Posted by slug58 on Fri, 19 Jul 2019 05:07:15 +0200

HBase Phoenix Helps Real-time Analysis of Massive Data

Preface Phoenix is a project that I haven't had much contact with. When a company I met last year used Phoenix to analyze tens of billions of records and delay in returning them at the second level, I slowly explored the inside of some phoenix. Last week, I talked with a Phoenix PMC & Committer about the location and future development of P ...

Posted by akreation on Sat, 29 Jun 2019 22:59:41 +0200

Hbase cluster deployment

Hhase cluster deployment Software used hadoop-2.7.4 hbase-1.2.6 jdk-8u144 zookeeper-3.4.10 Hbase comes with zookeeper, where you use your own deployed zookeeper zookeeper cluster deployment Install jdk Download zookeeper program Modify zoo.cfg tickTime=2000 initLimit=10 syncLimit=5 dataLogDir=/zookeeper/logs dataDir=/zookeeper ...

Posted by jebadoa on Wed, 29 May 2019 21:53:18 +0200

Big Data Learning Series 5 - Hive Integrating HBase Graphics and Texts

http://www.cnblogs.com/xuwujing/p/8059079.html Introduction In the last article Big Data Learning Series IV - --- Hadoop+Hive Environment Construction Graphics and Text Details (stand-alone) And before Big Data Learning Series II - HBase Environment Construction (stand-alone) Hive and HBase environments were successfully built and teste ...

Posted by dirkbonenkamp on Sat, 18 May 2019 16:02:04 +0200