Hadoop HBase Serialization and DeSerialization Tutorial, are you looking for the information of Serialization and DeSerialization in hadoop? Or the one who is casually glancing for the best platform which is providing bucketing in hive with examples for beginners? Then you’ve landed on the Right Platform which is packed with tons of Tutorials of bucketing in Hadoop hive. Follow the below mentioned Bucketing in Hive tutorial for Beginners which were originally designed by the world-class Trainers of Hadoop Training institute Professionals.
If you are the one who is a hunger to become the certified Pro Hadoop Developer? Or the one who is looking for the best Hadoop Training institute in Bangalore which offering advanced tutorials and Hadoop certification course to all the tech enthusiasts who are eager to learn the technology from starting Level to Advanced Level.
avro-1.8.1.jar
avro-tools-1.8.1.jar
log4j-api-2.0-beta9.jar
log4j-core-2.0-beta9.jar
$ cd /home/cloudera/Desktop/Avro
$ mkdir /home/hduser/Desktop/AVRO/schema
$ cd /home/cloudera/Desktop/Avro/Schema
$ gedit Prwatech.avsc
>> Open eclipse and write following program:
package com.test.avro;
import java.io.File;
import java.io.IOException;
import org.apache.avro.Schema;
import org.apache.avro.file.DataFileWriter;
import org.apache.avro.generic.GenericData;
import org.apache.avro.generic.GenericDatumWriter;
import org.apache.avro.generic.GenericRecord;
import org.apache.avro.io.DatumWriter;
public class SerializeNew {
public static void main(String args[]) throws IOException {
// Instantiating the Schema.Parser class.
Schema schema = new Schema.Parser().parse(new File(
“/home/cloudera/Desktop/Avro/Schema/Prwatech.avsc”));
// Instantiating the GenericRecord class.
GenericRecord e1 = new GenericData.Record(schema);
// Insert data according to schema
e1.put(“name”, “Ramman”);
e1.put(“id”, 001);
e1.put(“course”, “Hadoop”);
e1.put(“DOQ”, “Feb”);
e1.put(“location”, “Bangalore”);
GenericRecord e2 = new GenericData.Record(schema);
e2.put(“name”, “Akash”);
e2.put(“id”, 002);
e2.put(“course”, “Python”);
e2.put(“DOQ”, “March”);
e2.put(“location”, “Pune”);
DatumWriter<GenericRecord> datumWriter = new GenericDatumWriter<GenericRecord>( schema);
DataFileWriter<GenericRecord> dataFileWriter = new DataFileWriter<GenericRecord>( datumWriter);
dataFileWriter.create(schema, new File( “/home/cloudera/Desktop/Avro/data.txt”));
dataFileWriter.append(e1);
dataFileWriter.append(e2);
dataFileWriter.close();
System.out.println(“data successfully serialized”);
}
}
package com.test.avro;
import java.io.File;
import org.apache.avro.Schema;
import org.apache.avro.file.DataFileReader;
import org.apache.avro.generic.GenericDatumReader;
import org.apache.avro.generic.GenericRecord;
import org.apache.avro.io.DatumReader;
public class DeserializeNew {
public static void main(String args[]) throws Exception {
// Instantiating the Schema.Parser class.
Schema schema = new Schema.Parser().parse(new File(
/home/cloudera/Desktop/Avro/Schema/Prwatech.avsc”));
DatumReader<GenericRecord> datumReader = new GenericDatumReader<GenericRecord>(schema);
DataFileReader<GenericRecord> dataFileReader = new DataFileReader<GenericRecord>( new File(“/home/cloudera/Desktop/Avro/data.txt”),datumReader);
GenericRecord emp = null;
while (dataFileReader.hasNext()) {
emp = dataFileReader.next(emp);
System.out.println(emp);
}
System.out.println(“data deserialized”);
}
}
Get success in your career as a Hadoop developer by being a part of the Prwatech, India’s leading Hadoop training institute in Bangalore.