You are on page 1of 3

package org.bigdata.

mr;

import java.io.IOException;

import org.apache.hadoop.conf.Configuration;

import org.apache.hadoop.fs.FileSystem;

import org.apache.hadoop.fs.Path;

import org.apache.hadoop.io.IntWritable;

import org.apache.hadoop.io.Text;

import org.apache.hadoop.mapreduce.Job;

import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;

import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;

import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;

import org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat;

//import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;

public class TotalScoreDriver {

public static void main(String[] args) throws IOException, ClassNotFoundException,


InterruptedException {

Job myjob = new Job();

myjob.setJobName("TotalScores MR");

myjob.setJarByClass(TotalScoreDriver.class);

myjob.setMapperClass(TotalScoreMapper.class);

myjob.setReducerClass(TotalScoreReducer.class);

//myjob.setCombinerClass(TotalScoreReducer.class);

myjob.setPartitionerClass(MyPartitioner.class);
myjob.setMapOutputKeyClass(Text.class);

myjob.setMapOutputValueClass(IntWritable.class);

myjob.setOutputKeyClass(Text.class);

myjob.setOutputValueClass(IntWritable.class);

FileInputFormat.addInputPath(myjob, new Path(args[0]));

FileOutputFormat.setOutputPath(myjob, new Path(args[1]));

myjob.setInputFormatClass(TextInputFormat.class);

//myjob.setOutputFormatClass(TextOutputFormat.class);

myjob.setOutputFormatClass(SequenceFileOutputFormat.class);

//Set the desired number of reducers

//myjob.setNumReduceTasks(4);

//delete output directory if exists

FileSystem hdfs = FileSystem.get(new Configuration());

hdfs.delete(new Path(args[1]),true);

//Submit the Job

myjob.waitForCompletion(true);
}

You might also like