How to move data from Oracle database to Hadoop

0 votes
I've around 50 GB of data and my requirement is to import that data from Oracle database to Hadoop Distributed File System so that I can do further processing on that data.

So, can anyone tell me is there any way to load data from Oracle DB to HDFS?

I heard about a tool called Sqoop, but I don't have any idea how to use it?

Please help!

Thanks in advance.
Apr 12, 2018 in Big Data Hadoop by Shubham
• 13,490 points
7,337 views

1 answer to this question.

0 votes

Yes, you heard it correctly.

Apache Sqoop is one of the tool used for loading data from Oracle DB to HDFS. But, Sqoop is not just limited to just Oracle DB, you can use it to import and export data to & from most of the relational databases.

Let me tell you first about Sqoop.

Apache Sqoop is a tool designed for efficiently transferring bulk data between Apache Hadoop and external data stores such as relational databases, enterprise data warehouses.

Sqoop is used to import data from external datastores into Hadoop Distributed File System or related Hadoop eco-systems like Hive and HBase. Similarly, Sqoop can also be used to extract data from Hadoop or its eco-systems and export it to external datastores such as relational databases, enterprise data warehouses. Sqoop works with relational databases such as Teradata, Netezza, Oracle, MySQL, Postgres etc.

You can refer the link: https://www.edureka.co/blog/hdfs-using-sqoop/

Follow the steps as mentioned in the blog and you will get to know how Sqoop is used to transfer data from Oracle DB to HDFS.

I hope this will answer your question to some extent.

answered Apr 12, 2018 by nitinrawat895
• 11,380 points
how about the visa versa

transfer data from HDFS to Oracle?

do you have any practical example?

To transfer data from HDFS to Oracle you can use HIVE . This allows you to connect your database with HDFS.

You can go through the below link and end of the blog you will find similar kind of example.

https://www.edureka.co/blog/hive-commands-with-examples

Related Questions In Big Data Hadoop

0 votes
1 answer

How to fetch more data from database while using Sqoop import?

Hey. You guessed it right. The max ...READ MORE

answered May 15, 2019 in Big Data Hadoop by Gani
2,603 views
0 votes
1 answer

What are the different ways to load data from Hadoop to Azure Data Lake?

I would recommend you to go through ...READ MORE

answered Apr 18, 2018 in Big Data Hadoop by coldcode
• 2,090 points
1,073 views
0 votes
1 answer

How to import data to HBase from SQL server?

You can easily import the data from ...READ MORE

answered Apr 20, 2018 in Big Data Hadoop by kurt_cobain
• 9,350 points
1,648 views
0 votes
1 answer

How to implement data locality in Hadoop MapReduce?

You can use this getFileBlockLocations method of ...READ MORE

answered Apr 20, 2018 in Big Data Hadoop by kurt_cobain
• 9,350 points
1,124 views
–1 vote
1 answer

Hadoop dfs -ls command?

In your case there is no difference ...READ MORE

answered Mar 16, 2018 in Big Data Hadoop by kurt_cobain
• 9,350 points
4,642 views
+1 vote
1 answer

Hadoop Mapreduce word count Program

Firstly you need to understand the concept ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 11,380 points
11,077 views
+2 votes
11 answers

hadoop fs -put command?

Hi, You can create one directory in HDFS ...READ MORE

answered Mar 16, 2018 in Big Data Hadoop by nitinrawat895
• 11,380 points
109,073 views
0 votes
1 answer
0 votes
1 answer

How can we send data from MongoDB to Hadoop?

The MongoDB Connector for Hadoop reads data ...READ MORE

answered Mar 27, 2018 in Big Data Hadoop by nitinrawat895
• 11,380 points
2,120 views
0 votes
1 answer

How to extract only few lines of data from HDFS?

Here also in case of Hadoop, it is ...READ MORE

answered May 2, 2018 in Big Data Hadoop by nitinrawat895
• 11,380 points
11,567 views
webinar REGISTER FOR FREE WEBINAR X
REGISTER NOW
webinar_success Thank you for registering Join Edureka Meetup community for 100+ Free Webinars each month JOIN MEETUP GROUP