How does Hadoop accesses the files which are distributed among different boundaries?

0 votes
May 7 in Big Data Hadoop by nitinrawat895
• 9,030 points
10 views

1 answer to this question.

Your answer

Your name to display (optional):
Privacy: Your email address will only be used for sending these notifications.
0 votes

Hadoop's MapReduce function does not work on physical blocks of the file, instead, it is designed to work upon the logical memory or in simpler words, the input splits. 

These Input splits are dependent on the location where the file is written. A record may map two mappers.

The HDFS is designed in such a way that each and every file is written into it is split into blocks of 128 MB each and each block is replicated 3 times by default.

for example, consider a file. The data in this file can begin in block a and end in block b.

HDFS does not track the location of the data. Instead, it solely depends upon the logical input splits. It is these input splits which depict the start and end of any particular file.

enter image description here

for more information, you can go through this article.

answered May 7 by ravikiran
• 1,460 points

Related Questions In Big Data Hadoop

0 votes
1 answer

How can Hadoop process the records that are split across the block boundaries?

First of all, Map Reduce algorithm is not programmed ...READ MORE

answered Apr 15 in Big Data Hadoop by nitinrawat895
• 9,030 points
53 views
0 votes
1 answer

What are the different ways to load data from Hadoop to Azure Data Lake?

I would recommend you to go through ...READ MORE

answered Apr 18, 2018 in Big Data Hadoop by coldcode
• 1,980 points
41 views
0 votes
1 answer
0 votes
1 answer
0 votes
1 answer

Hadoop Mapreduce word count Program

Firstly you need to understand the concept ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 9,030 points
1,634 views
0 votes
1 answer

hadoop.mapred vs hadoop.mapreduce?

org.apache.hadoop.mapred is the Old API  org.apache.hadoop.mapreduce is the ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 9,030 points
130 views
0 votes
10 answers

hadoop fs -put command?

copy command can be used to copy files ...READ MORE

answered Dec 7, 2018 in Big Data Hadoop by Sujay
7,926 views
0 votes
1 answer

Hadoop dfs -ls command?

In your case there is no difference ...READ MORE

answered Mar 16, 2018 in Big Data Hadoop by kurt_cobain
• 9,260 points
549 views
0 votes
1 answer
0 votes
1 answer

Explain to me the method to transfer data between Azure tables and Hadoop on Azure

this article on HiveStorageHandler will let you create ...READ MORE

answered May 2 in Big Data Hadoop by ravikiran
• 1,460 points
14 views

© 2018 Brain4ce Education Solutions Pvt. Ltd. All rights Reserved.
"PMP®","PMI®", "PMI-ACP®" and "PMBOK®" are registered marks of the Project Management Institute, Inc. MongoDB®, Mongo and the leaf logo are the registered trademarks of MongoDB, Inc.