How to find the number of records using Map Reduce

How to find the number of records using Map Reduce


In real time scenario, data files contain many records. Also, there may be many data files available. In that case, it’s good to find a suitable approach to find out the output. Here, we want total number of records available in data files. So the requirement is to how to find the number of records using Map Reduce.

Components Involved

  • HDFS: Here it is used to store the input data which will get pass to the Map Reduce as an input and also store the map reduce output.
  • Map Reduce: Here it is used to process the data/file available in HDFS and store the output to the HDFS.

Sample Data

Let’s have a look at the sample data, how it will look like:

  1. id,first_name,last_name,gender,designation,city,country
  2. 1,Thayne,Mocher,Male,Administrative Officer,Port Colborne,Canada
  3. 2,Shelly,Bulfoot,Female,Analyst Programmer,Bacong,Philippines
  4. 3,Hercule,Chorlton,Male,Operator,Al Mazzunah,Tunisia
  5. 4,Thorstein,Epton,Male,Administrative Officer,Tayirove,Ukraine
  6. 5,Madelena,Savin,Female,Help Desk Technician,Jinjiang,China
  7. 6,Adeline,Imesson,Female,Legal Assistant,Fort Beaufort,South Africa
  8. 7,Celie,Richards,Male,Chemical Engineer,Dubiecko,Poland
  9. 8,Lilas,Harrowing,Female,Assistant Media Planner,Guayata,Colombia
  10. 9,Freida,Leivers,Female,Legal Assistant,Bangus Kulon,Indonesia
  11. 10,Celie,Dolligon,Female,Data Coordiator,Paraty,Brazil
  12. 11,Berkley,Orteaux,Male,Assistant Professor,Zinder,Niger
  13. 12,Gilburt,Minot,Male,Marketing Assistant,Hanyuan,China
  14. 13,Blaine,Treverton,Male,Research Associate,Yuankeng,China
  15. 14,Benjamen,Dodd,Male,Assistant Professor,Beberon,Philippines
  16. 15,Nikos,Worpole,Male,Human Resources Assistant II,Esmeralda,Cuba
  17. 16,Hercule,Richards,Male,Chemical Engineer,Dubiecko,Poland

Here, the sample data contains user’s info which is a comma(,) separated. You can download the sample data file from here.



The solution has many steps from setting up to the output of the code to the execution and validates the output. Let’s go ahead one by one:

Step 1: Input Data Preparation

The first step includes data preparation. In real time scenario, you will have your data. The data preparation means that it should present at the location from where our map reduce job can get it as an input file.

Let’s keep the sample data file at the local path. In my case, file name is sampledataForDuplicate” and local path is /home/NN/HadoopRepo/MapReduce/resources/recordCount”.

We need to move this data file to an HDFS location which will be input path for the map reduce.

hadoop fs -put /home/NN/HadoopRepo/MapReduce/resources/recordCount /user/bdp/mapreduce

It will copy the data file from local to HDFS location.

Step 2: Create Maven Project

In order to write map reduce program, create a maven project. We will add the dependency for the required package.

Follow the below steps in order to create the maven project:

  • Open Eclipse
  • Create a maven project

Step 3: Resolve Dependency

Add below dependency in pom.XML file and resolve dependency using cmd:

<!-- Hadoop Mapreduce Client Core -->

Step 4: Write Mapper

Once you are done with all above steps, write a mapper class which will take an input file. It will read the file and store each word of the file with key-value pair. Here using a java program to write the mapper.

package com.bdp.mapreduce.recordcount.mapper;
import org.apache.hadoop.mapreduce.Mapper;
public class RecordCountMapper 
        extends Mapper<LongWritable, Text, Text, IntWritable>{
        private static final IntWritable  one = new IntWritable(1);
        private Text record = new Text("Record");
        protected void map(LongWritable key, Text value,
                Mapper<LongWritable, Text, Text, IntWritable>.Context context)
                throws IOException, InterruptedException {
                // TODO Auto-generated method stub
                // Directly writing values which is nothing but a record
                if (key.get() == 0 && value.toString().contains("first_name")) {
                } else 
                        context.write(record, one);

Step 5: Write Reducer

In this step, we will take the mapper data as an input and process it. The actual logic has been written here.

Find the code below:

package com.bdp.mapreduce.recordcount.reducer;
import org.apache.hadoop.mapreduce.Reducer;
public class RecordCountReducer 
      extends Reducer<Text, IntWritable, NullWritable, IntWritable>{
      protected void reduce(Text key, Iterable<IntWritable> values,
            Reducer<Text, IntWritable, NullWritable, IntWritable>.Context context)
            throws IOException, InterruptedException {
            // TODO Auto-generated method stub
            int recordCount = 0;
            for (IntWritable value : values) {
                  recordCount += value.get();
            context.write( NullWritable.get(), new IntWritable(recordCount));

Step 6: Write Driver

In order to execute the mapper and reducer, let’s create a driver class which will call mapper and reducer. Find below the driver class code:

package com.bdp.mapreduce.recordcount.driver;
import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;
import com.bdp.mapreduce.recordcount.mapper.RecordCountMapper;
import com.bdp.mapreduce.recordcount.reducer.RecordCountReducer;
public class RecordCountDriver 
        extends Configured implements Tool{
        public int run(String[] args) throws Exception {
                // TODO Auto-generated method stub
                Job job = new Job(getConf(), "Record Count");
                FileInputFormat.addInputPath(job, new Path(args[0]));
                FileOutputFormat.setOutputPath(job, new Path(args[1]));
                return job.waitForCompletion(true) ? 0 : 1;      
        public static void main(String[] args) throws Exception {
                int jobStatus = RecordCountDriver(), args);

Step 7: Package Preparation

In this step, we will create a package(.jar) of the project. Follow the below steps:

  • Open CMD
  • Head to the maven project
  • Use the command: mvn package

You can also download the package which I have built for this requirement:


Step 8: Execution

All the setup has been done. Let’s execute the job and validate the output. In order to execute the map reduce, use below command:
Format: hadoop jar <path of jar> <driver class with package name> <input data path of HDFS> <output path at HDFS>

hadoop jar /home/NN/HadoopRepo/MapReduce/MapReduceForRecordCount-0.0.1-SNAPSHOT.jar com.bdp.mapreduce.recordcount.driver.RecordCountDriver /user/bdp/mapreduce/recordCount /user/bdp/mapreduce/out/recordCount

Step 9: Validate Output

Check the output at HDFS path.

[root@NN hadoop-2.6.0]# hadoop fs -ls /user/bdp/mapreduce/out/recordCount
17/08/01 22:04:49 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Found 2 items
-rw-r--r--   1 root supergroup          0 2017-08-01 22:04 /user/bdp/mapreduce/out/recordCount/_SUCCESS
-rw-r--r--   1 root supergroup          3 2017-08-01 22:04 /user/bdp/mapreduce/out/recordCount/part-r-00000
print output
[root@NN hadoop-2.6.0]# hadoop fs -cat /user/bdp/mapreduce/out/recordCount/part-r-00000
17/08/01 22:05:11 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable

Wrapping Up

In this post, we have written a map reduce to count the number of records of the data file(s). This is similar to SQL count(*) command. Here, this MapReduce job contains three class map, reduce and driver. Mapper reads the input, Reducer counts the records and Driver running setting all the configuration to run the map reduce.


Join in hive with example

Requirement You have two table named as A and B. and you want to perform all types of join in ...
Read More

Join in pyspark with example

Requirement You have two table named as A and B. and you want to perform all types of join in ...
Read More

Join in spark using scala with example

Requirement You have two table named as A and B. and you want to perform all types of join in ...
Read More

Java UDF to convert String to date in PIG

About Code Many times it happens like you have received data from many systems and each system operates on a ...
Read More
/ java udf, Pig, pig, pig udf, string to date, udf

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.