Integrating data from multiple sources is essential in the age of big data, but it can be a challenging and time-consuming task. This handy cookbook provides dozens of ready-to-use recipes for using Apache Sqoop, the command-line interface application that optimizes data transfers between relational databases and Hadoop.
Sqoop is both powerful and bewildering, but with this cookbook’s problem-solution-discussion format, you’ll quickly learn how to deploy and then apply Sqoop in your environment. The authors provide MySQL, Oracle, and PostgreSQL database examples on GitHub that you can easily adapt for SQL Server, Netezza, Teradata, or other relational systems.
Transfer data from a single database table into your Hadoop ecosystem
Keep table data and Hadoop in sync by importing data incrementally
Import data from more than one database table
Customize transferred data by calling various database functions
Export generated, processed, or backed-up data from Hadoop to your database
Run Sqoop within Oozie, Hadoop’s specialized workflow scheduler
Load data into Hadoop’s data warehouse (Hive) or database (HBase)
Handle installation, connection, and syntax issues common to specific database vendors
評分
評分
評分
評分
小巧實用,簡明易讀
评分要是能有API編程的相關內容就更好瞭
评分一問一答得方式解決問題,十分簡短,個人覺得相當不錯。
评分工具書
评分小巧實用,簡明易讀
本站所有內容均為互聯網搜索引擎提供的公開搜索信息,本站不存儲任何數據與內容,任何內容與數據均與本站無關,如有需要請聯繫相關搜索引擎包括但不限於百度,google,bing,sogou 等
© 2025 book.quotespace.org All Rights Reserved. 小美書屋 版权所有