This is documentation for MapR Version 5.0. You can also refer to MapR documentation for the latest release.

Skip to end of metadata
Go to start of metadata

Apache Sqoop™ is a tool designed for efficiently transferring bulk data between Apache Hadoop and structured datastores such as relational databases.

This documentation provides all relevant details about using Sqoop and Sqoop2 with MapR, but does not duplicate Apache documentation. You can refer also to documentation available from the Apache Sqoop website

The following table describes the differences between Sqoop1 or Sqoop2:

FeatureSqoop1Sqoop2
Specialized connectors for all major RDBMSAvailable.

Not available. However, you can use the generic JDBC connector, which has been tested on these databases:

  • MySQL
  • Microsoft SQL Server
  • Oracle
  • PostgreSQL

The generic JDBC connector should also work with any other JDBC-compliant database, although specialized connectors probably give better performance.

Data transfer from RDBMS to Hive or HBaseDone automatically.

Must be done manually in two stages:

  1. Import data from RDBMS into MapR-FS.
  2. Load data into Hive (using the LOAD DATA command) or HBase
Data transfer from Hive or HBase to RDBMS

Must be done manually in two stages:

  1. Extract data from Hive or HBase into MapR-FS, as a text file or as an Avro file.
  2. Export the output of step 1 to an RDBMS using Sqoop.

Must be done manually in two stages:

  1. Extract data from Hive or HBase into MapR-FS, as a text file or as an Avro file.
  2. Export the output of step 1 to an RDBMS using Sqoop.
Integrated Kerberos securitySupported.Not supported.
Password encryptionNot supported.Supported using Derby's data encryption feature (although the configuration has not been verified).
  • No labels