0% found this document useful (0 votes)
31 views3 pages

Hadoop Word Count Example

The document contains a Java program for a Hadoop MapReduce application that performs word count. It includes a Mapper class to tokenize input text and a Reducer class to sum the occurrences of each word. The main method sets up the job configuration and specifies input and output paths.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
31 views3 pages

Hadoop Word Count Example

The document contains a Java program for a Hadoop MapReduce application that performs word count. It includes a Mapper class to tokenize input text and a Reducer class to sum the occurrences of each word. The main method sets up the job configuration and specifies input and output paths.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd

Name: Kadam Arati Dilip

Roll No:COTA29
Code:
import java.io.IOException; import java.util.StringTokenizer; import

org.apache.hadoop.conf.Configuration; import

org.apache.hadoop.fs.Path; import

org.apache.hadoop.io.IntWritable; import

org.apache.hadoop.io.Text; import

org.apache.hadoop.mapreduce.Job; import

org.apache.hadoop.mapreduce.Mapper; import

org.apache.hadoop.mapreduce.Reducer; import

org.apache.hadoop.mapreduce.lib.input.FileInputFormat; import

org.apache.hadoop.mapreduce.lib.output.FileOutputFormat; public

class WordCount {

public static class TokenizerMapper extends

Mapper<Object, Text, Text, IntWritable>{

private final static IntWritable one = new IntWritable(1);

private Text word = new Text();

public void map(Object key, Text value, Context context

) throws IOException, InterruptedException

{ StringTokenizer itr = new StringTokenizer(value.toString());

while (itr.hasMoreTokens()) { word.set(itr.nextToken());

context.write(word, one);

} }}
public static class IntSumReducer extends

Reducer<Text,IntWritable,Text,IntWritable> { private

IntWritable result = new IntWritable();

public void reduce(Text key, Iterable<IntWritable>

values, Context context

) throws IOException, InterruptedException {

int sum = 0; for (IntWritable val : values) { sum +=

val.get();

result.set(sum);

context.write(key, result);

public static void main(String[] args) throws Exception {

Configuration conf = new Configuration(); Job job =

Job.getInstance(conf, "word count");

job.setJarByClass(WordCount.class);

job.setMapperClass(TokenizerMapper.class);

job.setCombinerClass(IntSumReducer.class);

job.setReducerClass(IntSumReducer.class);

job.setOutputKeyClass(Text.class);

job.setOutputValueClass(IntWritable.class);

FileInputFormat.addInputPath(job, new Path(args[0]));

FileOutputFormat.setOutputPath(job, new Path(args[1]));

System.exit(job.waitForCompletion(true) ? 0 : 1);

}
Input:
Aditya

Onkar

Nangare

Nimse

Onkar

Shubham

Ganesh

Aditya

Output:

You might also like