Integrating data from multiple sources is essential in the age of big data but it can be a challenging & time-consuming task This handy cookbook provides dozens of ready-to-use recipes for using Apache Sqoop the comm&-line interface application that optimizes data transfers between relational databases & Hadoop Sqoop is both powerful & bewildering but with this cookbook's problem-solution-discussion format you'll quickly learn how to deploy & then apply Sqoop in your environment The authors provide My SQL Oracle & Postgre SQL database examples on Git Hub that you can easily adapt for SQL Server Netezza Teradata or other relational systems Transfer data from a single database table into your Hadoop ecosystem Keep table data & Hadoop in sync by importing data incrementally Import data from more than one database table Customize transferred data by calling various database functions Export generated processed or backed-up data from Hadoop to your database Run Sqoop within Oozie Hadoop's specialized workflow scheduler Load data into Hadoop's data warehouse (Hive) or database (HBase) Handle installation connection & syntax issues common to specific database vendors