Load data into teradata using sqoop export

0 votes
Hi Team,

I have one partition table, so how i can load data into teradata using sqoop export.
Aug 9 in Big Data Hadoop by Prateek
83 views

1 answer to this question.

0 votes

The general syntax to do this as shown below.

hive table:

CREATE EXTERNAL TABLE mytab_load(field1 STRING, field2 STRING, field3 STRING)

ROW FORMAT DELIMITED 

FIELDS TERMINATED BY '\t' 

STORED AS TEXTFILE 

LOCATION '/maprfs;/../../mytab_load' 

Sqoop Export:

sqoop export --driver com.teradata.jdbc.TeraDriver --connect "jdbc:teradata://localhost/database=dbname" --password "xxxx" --username "usser" --table "mytab_load" --num-mappers 1 --export-dir "/maprfs:/../../mytab_load" --input-fields-terminated-by '\t' -Dsqoop.export.records.per.statement=1 
answered Aug 9 by Kunal

Related Questions In Big Data Hadoop

0 votes
1 answer

Load data into teradata using sqoop

The general syntax to do this as ...READ MORE

answered Jul 15 in Big Data Hadoop by Ritu
56 views
0 votes
1 answer

How to transfer data from Netezza to HDFS using Apache Sqoop?

Remove the --direct option. It gives issue ...READ MORE

answered Apr 23, 2018 in Big Data Hadoop by kurt_cobain
• 9,260 points
315 views
0 votes
1 answer

Hadoop Hive Hbase: How to insert data into Hbase using Hive (JSON file)?

You can use the get_json_object function to parse the ...READ MORE

answered Nov 15, 2018 in Big Data Hadoop by Omkar
• 67,620 points
582 views
+1 vote
1 answer

What is the process to perform an incremental data load in Sqoop?

The process to perform incremental data load ...READ MORE

answered Dec 17, 2018 in Big Data Hadoop by Frankie
• 9,810 points
425 views
0 votes
1 answer

Hadoop Mapreduce word count Program

Firstly you need to understand the concept ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 10,690 points
3,024 views
0 votes
1 answer

hadoop.mapred vs hadoop.mapreduce?

org.apache.hadoop.mapred is the Old API  org.apache.hadoop.mapreduce is the ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 10,690 points
337 views
0 votes
10 answers

hadoop fs -put command?

put syntax: put <localSrc> <dest> copy syntax: copyFr ...READ MORE

answered Dec 7, 2018 in Big Data Hadoop by Aditya
14,936 views
0 votes
1 answer

Hadoop dfs -ls command?

In your case there is no difference ...READ MORE

answered Mar 16, 2018 in Big Data Hadoop by kurt_cobain
• 9,260 points
1,109 views
+3 votes
1 answer

Getting Connection Error while loading data into table using cloudera hive

Hey Nafeesa, Itseems that Hive is not able ...READ MORE

answered Oct 3, 2018 in Big Data Hadoop by Vardhan
• 12,570 points
76 views
0 votes
1 answer

How to securely transfer data from rdms to hdfs using sqoop?

Sqoop stores metadata in a repository and ...READ MORE

answered Dec 18, 2018 in Big Data Hadoop by Omkar
• 67,620 points
94 views