Cantitate/Preț
Produs

Hadoop Application Architectures

Autor Mark Grover, Ted Malaska, Jonathan Seidman, Gwen Shapira
en Limba Engleză Paperback – 9 iul 2015
Get expert guidance on architecting end-to-end data management solutions with Apache Hadoop. While many sources explain how to use various components in the Hadoop ecosystem, this practical book takes you through architectural considerations necessary to tie those components together into a complete tailored application, based on your particular use case.

To reinforce those lessons, the book’s second section provides detailed examples of architecture used in some of the most commonly found Hadoop applications. Whether you’re designing and implementing a new Hadoop application, or planning to integrate Hadoop into your existing data infrastructure, Hadoop Application Architectures will skillfully guide you through the process.
Citește tot Restrânge

Preț: 28493 lei

Preț vechi: 35616 lei
-20% Nou

Puncte Express: 427

Preț estimativ în valută:
5453 5753$ 4544£

Carte tipărită la comandă

Livrare economică 02-16 ianuarie 25

Preluare comenzi: 021 569.72.76

Specificații

ISBN-13: 9781491900086
ISBN-10: 1491900083
Pagini: 400
Ilustrații: black & white illustrations
Dimensiuni: 179 x 234 x 25 mm
Greutate: 0.64 kg
Editura: O'Reilly

Descriere

Get expert guidance on architecting end-to-end data management solutions with Apache Hadoop. While many sources explain how to use various components in the Hadoop ecosystem, this practical book takes you through architectural considerations necessary to tie those components together into a complete tailored application, based on your particular use case. To reinforce those lessons, the book's second section provides detailed examples of architectures used in some of the most commonly found Hadoop applications. Whether you're designing a new Hadoop application, or planning to integrate Hadoop into your existing data infrastructure, Hadoop Application Architectures will skillfully guide you through the process. This book covers: Factors to consider when using Hadoop to store and model data Best practices for moving data in and out of the system Data processing frameworks, including MapReduce, Spark, and Hive Common Hadoop processing patterns, such as removing duplicate records and using windowing analytics Giraph, GraphX, and other tools for large graph processing on Hadoop Using workflow orchestration and scheduling tools such as Apache Oozie Near-real-time stream processing with Apache Storm, Apache Spark Streaming, and Apache Flume Architecture examples for clickstream analysis, fraud detection, and data warehousing

Notă biografică