MapReduce Fundamentals - BunksAllowed

BunksAllowed is an effort to facilitate Self Learning process through the provision of quality tutorials.

Random Posts

MapReduce Fundamentals

Share This

What is MapReduce?

MapReduce is a programming model and processing technique designed for processing and generating large datasets in a parallel and distributed fashion. It was introduced by Google and popularized by Apache Hadoop, an open-source framework. MapReduce divides a computation task into two phases: the Map phase and the Reduce phase.

MapReduce is an open-source software framework for distributed storage and processing of large datasets across clusters of computers using simple programming models. MapReduce in Hadoop refers to the implementation of the MapReduce programming model within the Apache Hadoop framework. MapReduce is the primary processing engine in Hadoop's implementation, enabling users to create distributed applications that handle enormous volumes of data in parallel over a distributed cluster. The framework effectively manages fault tolerance and data locality by distributing data and computation over several cluster nodes. The two main stages of Hadoop's MapReduce process are the "map" phase, which splits input data into smaller pieces and processes them concurrently across several nodes, and the "reduce" phase, which gathers and processes the intermediate results produced by the map phase to create the final output. Hadoop is an essential part of big data analytics and processing pipelines because of its distributed processing approach, which makes it possible for it to handle and analyze large-scale datasets with efficiency.

Key Concepts:


  • Map Function: Processes input data and produces a set of intermediate key-value pairs.
  • Shuffling and Sorting: The intermediate key-value pairs are shuffled and sorted based on keys.
  • Reduce Function: Takes the sorted key-value pairs, groups them by key, and performs a specified operation on each group.

JobTracker and TaskTracker 

 
In Hadoop MapReduce, the JobTracker and TaskTracker play crucial roles in managing and executing MapReduce jobs within a Hadoop cluster. The JobTracker is responsible for coordinating and managing MapReduce jobs submitted by users. It is typically a single master node in the Hadoop cluster. 
 
Its main functions include:
 
The JobTracker schedules MapReduce jobs for execution based on the available cluster resources and job priorities.
  • Task Assignment: It allots map and reduce jobs to the cluster's available TaskTrackers.
  • Monitoring: The JobTracker tracks the success and failure of individual tasks as well as the progress of MapReduce jobs.
  • Fault Tolerance: TaskTracker or individual task failures are detected and handled to guarantee fault tolerance. A JobTracker transfers its tasks to other TaskTrackers that are available in the event that a TaskTracker fails.
The TaskTrackers, on the other hand, are slave nodes in the Hadoop cluster that carry out the map and reduce tasks that the JobTracker assigns them. Among their principal roles are:
  • Task Execution: TaskTrackers carry out the tasks that the JobTracker assigns them, including mapping and reducing. They handle the actual data processing.
  • Heartbeat: To let the JobTracker know they are still alive and well, TaskTrackers periodically send heartbeat signals to it. This makes it possible for the JobTracker to identify errors fast.
  • Task Progress Reporting: TaskTrackers send information about the success or failure of a task to the JobTracker by reporting task progress.
  • Speculative Execution: TaskTrackers have the ability to execute tasks speculatively in the event that they notice that a specific job is moving more slowly than anticipated on another TaskTracker. This reduces the time it takes to finish a project by executing duplicate jobs concurrently.
Together, JobTracker and TaskTracker ensure scalability, fault tolerance, and optimal resource use by effectively managing and executing MapReduce jobs in a distributed Hadoop environment.

Steps in Map Reduce 

  • The map takes data in the form of pairs and returns a list of <key, value> pairs. The keys will not be unique in this case. 
  • Using the output of Map, sort and shuffle are applied by the Hadoop architecture. This sort and shuffle acts on these list of <key, value> pairs and sends out unique keys and a list of values associated with this unique key <key, list(values)>. 
  • An output of sort and shuffle sent to the reducer phase. The reducer performs a defined function on a list of values for unique keys, and Final output <key, value> will be stored/displayed. 

Usage of MapReduce 

  • It can be used in various applications like document clustering, distributed sorting, and web link-graph reversal. 
  • It can be used for distributed pattern-based searching. 
  • We can also use MapReduce in machine learning. 
  • It was used by Google to regenerate Google's index of the World Wide Web. 
  • It can be used in multiple computing environments such as multi-cluster, multi-core, and mobile environments.

The MapReduce Process:


Map Phase:

Input Splitting: The input data is divided into fixed-size splits, and each split is processed by a separate map task.

Map Function Execution: The map function is applied to each record in the input split, generating a set of intermediate key-value pairs.

Intermediate Key-Value Pairs: The intermediate key-value pairs are temporarily stored in a data structure.

Shuffling and Sorting:

Partitioning: The intermediate key-value pairs are partitioned based on the keys.

Shuffle and Sort: Data is transferred over the network to the reducers. The shuffle and sort phase ensures that all values for a given key are sent to the same reducer.

Reduce Phase:

Grouping: The values for each key are grouped together.

Reduce Function Execution: The reduce function is applied to each group of values, producing the final output.

Output: The final output consists of the results of the reduce function applied to each key.


MapReduce Example:


Let's consider a simple word count example.

Map Function:
Input: (document_id, document_text)
def map_function(document_id, document_text):
	for word in document_text.split():
    	emit_intermediate(word, 1)

Reduce Function:
Input: (word, [1, 1, 1, ...])
def reduce_function(word, counts):
	total_count = sum(counts)
	emit_output(word, total_count)

Implementing MapReduce with Hadoop:


Writing Map and Reduce Programs:

Create map and reduce functions and save them in separate files, e.g., `map.py` and `reduce.py`.

Hadoop Streaming:

Use Hadoop Streaming to execute MapReduce jobs with custom scripts.

  hadoop jar $HADOOP_HOME/share/hadoop/tools/lib/hadoop-streaming-*.jar \
  		-file map.py -mapper map.py -file reduce.py \
        -reducer reduce.py -input input_dir -output output_dir
  

MapReduce is a powerful paradigm for processing large-scale data in a distributed and parallel manner. Understanding its key concepts, implementation steps, and optimization techniques is essential for leveraging its capabilities in the world of Big Data processing.

No comments:

Post a Comment