How to handle Sqoop failure program and restart the job from failed tables?

+3 votes
When running a Sqoop jobs on a large number of tables, the job fails at some table. I want to know how I can run the same Sqoop job again on only those tables that have failed without running the sqoop job all over again on all the tables.
Aug 29 in Big Data Hadoop by Kunal
• 150 points

edited Sep 5 by Kunal 51 views

No answer to this question. Be the first to respond.

Your answer

Your name to display (optional):
Privacy: Your email address will only be used for sending these notifications.

Related Questions In Big Data Hadoop

0 votes
1 answer

How to import tables into Hive from Sqoop?

Very simple, in your sqoop command that ...READ MORE

answered May 15 in Big Data Hadoop by John
41 views
0 votes
1 answer

How to create a FileSystem object that can be used for reading from and writing to HDFS?

Read operation on HDFS In order to read ...READ MORE

answered Mar 21, 2018 in Big Data Hadoop by nitinrawat895
• 10,670 points

edited Mar 21, 2018 by nitinrawat895 295 views
0 votes
11 answers
0 votes
1 answer

How to transfer data from Netezza to HDFS using Apache Sqoop?

Remove the --direct option. It gives issue ...READ MORE

answered Apr 23, 2018 in Big Data Hadoop by kurt_cobain
• 9,240 points
289 views
0 votes
1 answer
0 votes
1 answer

What is grunt shell in pig?

Hey, Grunt shell is a shell command.The Grunts ...READ MORE

answered May 6 in Big Data Hadoop by Gitika
• 25,340 points
40 views
0 votes
1 answer
0 votes
1 answer

Hadoop Mapreduce word count Program

Firstly you need to understand the concept ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 10,670 points
2,753 views
0 votes
1 answer

hadoop.mapred vs hadoop.mapreduce?

org.apache.hadoop.mapred is the Old API  org.apache.hadoop.mapreduce is the ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 10,670 points
289 views