Hadoop HBase Serialization and DeSerialization

  • date 28th April, 2019 |
  • by Prwatech |
  • 0 Comments

Hadoop HBase Serialization and DeSerialization Tutorial

 

Hadoop HBase Serialization and DeSerialization Tutorial, are you looking for the information of Serialization and DeSerialization in hadoop? Or the one who is casually glancing for the best platform which is providing bucketing in hive with examples for beginners? Then you’ve landed on the Right Platform which is packed with tons of Tutorials of bucketing in Hadoop hive. Follow the below mentioned Bucketing in Hive tutorial for Beginners which were originally designed by the world-class Trainers of Hadoop Training institute Professionals.

 

If you are the one who is a hunger to become the certified Pro Hadoop Developer? Or the one who is looking for the best Hadoop Training institute in Bangalore  which offering advanced tutorials and Hadoop certification course to all the tech enthusiasts who are eager to learn the technology from starting Level to Advanced Level.

 

HBase Configuration Serialization

 

Add these libraries to your java project build path.

 

avro-1.8.1.jar

avro-tools-1.8.1.jar

log4j-api-2.0-beta9.jar

log4j-core-2.0-beta9.jar

 

Step 1 – Change the directory to /home/cloudera/Desktop/Avro

 

$ cd /home/cloudera/Desktop/Avro

 

Step 2 – Make a new directory schema in /home/cloudera/Desktop/Avro

 

$ mkdir /home/hduser/Desktop/AVRO/schema

 

Step 3 – Change the directory to /home/cloudera/Desktop/Avro/Schema

 

$ cd /home/cloudera/Desktop/Avro/Schema

 

Step 4 – Create a new avro schema Prwatech.avsc in /home/cloudera/Desktop/Avro/schema. It creates a new Prwatech.avsc file if it doesn’t exists and opens for editing.

 

$ gedit Prwatech.avsc

 

Hadoop HBase Serialization and DeSerialization Tutorial

 

 

Step 5 – Add these following lines to Prwatech.avsc  file. Save and close it.

 

>> Open eclipse and write following program:

package com.test.avro;

import java.io.File;

import java.io.IOException;

import org.apache.avro.Schema;

import org.apache.avro.file.DataFileWriter;

import org.apache.avro.generic.GenericData;

import org.apache.avro.generic.GenericDatumWriter;

import org.apache.avro.generic.GenericRecord;

import org.apache.avro.io.DatumWriter;

 

public class SerializeNew {

public static void main(String args[]) throws IOException {

// Instantiating the Schema.Parser class.

Schema schema = new Schema.Parser().parse(new File(

“/home/cloudera/Desktop/Avro/Schema/Prwatech.avsc”));

// Instantiating the GenericRecord class.

GenericRecord e1 = new GenericData.Record(schema);

// Insert data according to schema

e1.put(“name”, “Ramman”);

e1.put(“id”, 001);

e1.put(“course”, “Hadoop”);

e1.put(“DOQ”, “Feb”);

e1.put(“location”, “Bangalore”);

GenericRecord e2 = new GenericData.Record(schema);

e2.put(“name”, “Akash”);

e2.put(“id”, 002);

e2.put(“course”, “Python”);

e2.put(“DOQ”, “March”);

e2.put(“location”, “Pune”);

DatumWriter<GenericRecord> datumWriter = new GenericDatumWriter<GenericRecord>( schema);

DataFileWriter<GenericRecord> dataFileWriter = new DataFileWriter<GenericRecord>( datumWriter);

dataFileWriter.create(schema, new File( “/home/cloudera/Desktop/Avro/data.txt”));

dataFileWriter.append(e1);

dataFileWriter.append(e2);

dataFileWriter.close();

System.out.println(“data successfully serialized”);

}

}

Now compile the program

 

 

Hadoop HBase Serialization and DeSerialization Tutorial

 

Output will be

 

Hadoop HBase Serialization and DeSerialization Tutorial

 

HBase Configuration DeSerialization

 

Input Data will be

 

Hadoop HBase Serialization and DeSerialization Tutorial

 

Run following command in eclipse within same package (but create different class)

 

package com.test.avro;

import java.io.File;

import org.apache.avro.Schema;

import org.apache.avro.file.DataFileReader;

import org.apache.avro.generic.GenericDatumReader;

import org.apache.avro.generic.GenericRecord;

import org.apache.avro.io.DatumReader;

public class DeserializeNew {

public static void main(String args[]) throws Exception {

// Instantiating the Schema.Parser class.

Schema schema = new Schema.Parser().parse(new File(

/home/cloudera/Desktop/Avro/Schema/Prwatech.avsc”));

DatumReader<GenericRecord> datumReader = new GenericDatumReader<GenericRecord>(schema);

DataFileReader<GenericRecord> dataFileReader = new DataFileReader<GenericRecord>( new File(“/home/cloudera/Desktop/Avro/data.txt”),datumReader);

GenericRecord emp = null;

while (dataFileReader.hasNext()) {

emp = dataFileReader.next(emp);

System.out.println(emp);

}

System.out.println(“data deserialized”);

}

}

 

After DeSerialization the output will be

 

 

Hadoop HBase Serialization and DeSerialization Tutorial

 

 

Get success in your career as a Hadoop developer by being a part of the Prwatech, India’s leading Hadoop training institute in Bangalore.

 

 

 

 

 

0
0

Quick Support

image image