Sqoop Export: Transferring Data from HDFS to RDBMS
Learn how to use Sqoop to export data from Hadoop Distributed File System (HDFS) to a relational database management system (RDBMS). This guide provides examples and best practices for exporting data using Sqoop.
Using Sqoop for Data Export
Exporting Data from HDFS to RDBMS
In previous examples, we used Sqoop to import data from a relational database management system (RDBMS) into Hadoop Distributed File System (HDFS). This section demonstrates how to use Sqoop's export
functionality to transfer data in the opposite direction—from HDFS to an RDBMS.
Prerequisites: Creating the RDBMS Table
Before exporting data, Sqoop needs information about the target table's structure in your RDBMS. You must create the table in your database beforehand. Here's an example using MySQL. Replace table_name
and column_name
/column_type
with your table and column details:
MySQL Command (Creating Table)
mysql> CREATE TABLE table_name (column_name column_type);
Exporting Data with Sqoop
Once the table is created, use the following Sqoop command to export data from HDFS to the database. Replace the connection details, username, table name, and HDFS path with your specific information:
Sqoop Export Command
sqoop export \
--connect jdbc:mysql://localhost/cloudera \
--username cloudera -P \
--table exported \
--export-dir /user/country_imported/part-m-00000
Verifying the Export
After running the export command, verify that the data has been successfully imported into your MySQL table:
MySQL Command (Verifying Data)
mysql> SELECT * FROM exported;
This demonstrates how to export data from HDFS back into a relational database using Sqoop's export functionality.