|
|
|
|
LEADER |
00000cam a22000007a 4500 |
001 |
EBOOKCENTRAL_ocn960040671 |
003 |
OCoLC |
005 |
20240329122006.0 |
006 |
m o d |
007 |
cr |n||||||||| |
008 |
161118s2016 xx o 000 0 eng d |
040 |
|
|
|a IDEBK
|b eng
|e pn
|c IDEBK
|d OCLCQ
|d IDEBK
|d COO
|d EBLCP
|d OCLCQ
|d MERUC
|d CHVBK
|d OCLCF
|d OCLCO
|d OCLCQ
|d OCLCO
|d LVT
|d UKAHL
|d OCLCQ
|d OCLCO
|d K6U
|d OCLCQ
|d OCLCO
|d OCLCL
|
020 |
|
|
|a 1783980311
|q (ebk)
|
020 |
|
|
|a 9781783980314
|
020 |
|
|
|z 1783980303
|
029 |
1 |
|
|a AU@
|b 000066230916
|
029 |
1 |
|
|a CHNEW
|b 000949170
|
029 |
1 |
|
|a CHVBK
|b 483153443
|
035 |
|
|
|a (OCoLC)960040671
|
037 |
|
|
|a 958873
|b MIL
|
050 |
|
4 |
|a T55.4-60.8
|
082 |
0 |
4 |
|a 005.8
|2 23
|
049 |
|
|
|a UAMI
|
100 |
1 |
|
|a Anurag Shrivastava; Tanmay Deshpande.
|
245 |
1 |
0 |
|a Hadoop Blueprints.
|
250 |
|
|
|a 1.
|
260 |
|
|
|b Packt Publishing,
|c 2016.
|
300 |
|
|
|a 1 online resource (316)
|
336 |
|
|
|a text
|b txt
|2 rdacontent
|
337 |
|
|
|a computer
|b c
|2 rdamedia
|
338 |
|
|
|a online resource
|b cr
|2 rdacarrier
|
588 |
0 |
|
|a Print version record.
|
505 |
0 |
|
|a Cover; Copyright; Credits; About the Authors; About the Reviewers; www.PacktPub.com; Table of Contents; Preface; Chapter 1: Hadoop and Big Data; The beginning of the big data problem; Limitations of RDBMS systems; Scaling out a database on Google; Parallel processing of large datasets; Building open source Hadoop; Enterprise Hadoop; Social media and mobile channels; Data storage cost reduction; Enterprise software vendors; Pure Play Hadoop vendors; Cloud Hadoop vendors; The design of the Hadoop system; The Hadoop Distributed File System (HDFS); Data organization in HDFS.
|
505 |
8 |
|
|a HDFS file management commandsNameNode and DataNodes; Metadata store in NameNode; Preventing a single point of failure with Hadoop HA; Checkpointing process; Data Store on a DataNode; Handshakes and heartbeats; MapReduce; The execution model of MapReduce Version 1; Apache YARN; Building a MapReduce Version 2 program; Problem statement; Solution workflow; Getting the dataset; Studying the dataset; Cleaning the dataset; Loading the dataset on the HDFS; Starting with a MapReduce program; Installing Eclipse; Creating a project in Eclipse; Coding and building a MapReduce program.
|
505 |
8 |
|
|a Run the MapReduce program locallyExamine the result; Run the MapReduce program on Hadoop; Further processing of results; Hadoop platform tools; Data ingestion tools; Data access tools; Monitoring tools; Data governance tools; Big data use cases; Creating a 360 degree view of a customer; Fraud detection systems for banks; Marketing campaign planning; Churn detection in telecom; Analyzing sensor data; Building a data lake; The architecture of Hadoop-based systems; Lambda architecture; Summary; Chapter 2: A 360-Degree View of the Customer; Capturing business information.
|
505 |
8 |
|
|a Collecting data from data sourcesCreating a data processing approach; Presenting the results; Setting up the technology stack; Tools used; Installing Hortonworks Sandbox; Creating user accounts; Exploring HUE; Exploring MYSQL and the HIVE command line; Exploring Sqoop at the command line; Test driving Hive and Sqoop; Querying data using Hive; Importing data in Hive using Sqoop; Engineering the solution; Datasets; Loading customer master data into Hadoop; Loading web logs into Hadoop; Loading tweets into Hadoop; Creating the 360-degree view; Exporting data from Hadoop; Presenting the view.
|
505 |
8 |
|
|a Building a web applicationInstalling Node.js; Coding the web application in Node.js; Summary; Chapter 3: Building a Fraud Detection System; Understanding the business problem; Selecting and cleansing the dataset; Finding relevant fields; Machine learning for fraud detection; Clustering as an unsupervised machine learning method; Designing the high-level architecture; Introducing Apache Spark; Apache Spark architecture; Resilient Distributed Datasets; Transformation functions; Actions; Test driving Apache Spark; Calculating the yearly average stock prices using Spark; Apache Spark 2.X.
|
520 |
|
|
|a About This BookSolve real-world business problems using Hadoop and other Big Data technologiesBuild efficient data lakes in Hadoop, and develop systems for various business cases like improving marketing campaigns, fraud detection, and morePower packed with six case studies to get you going with Hadoop for Business IntelligenceWho This Book Is For If you are interested in building efficient business solutions using Hadoop, this is the book for you This book assumes that you have basic knowledge of Hadoop, Java, and any scripting language. What You Will LearnLearn about the evolution of Hadoop as the big data platformUnderstand the basics of Hadoop architectureBuild a 360-degree view of your customer using Sqoop and HiveBuild and run classification models on Hadoop using BigMLUse Spark and Hadoop to build a fraud detection systemDevelop a churn detection system using Java and MapReduceBuild an IoT-based data collection and visualization systemGet to grips with building a Hadoopbased data lake for large enterprisesLearn about the coexistence of NoSQL and In-Memory databases in the Hadoop ecosystemIn Detail If you have a basic understanding of Hadoop and want to put your knowledge to use to build fantastic big data solutions for business, then this book is for you. Build six real-life, end-to-end solutions using the tools in the Hadoop ecosystem, and take your knowledge of Hadoop to the next level. Start off by understanding various business problems that can be solved using Hadoop. You will also get acquainted with the common architectural patterns that are used to build Hadoop-based solutions. Build a 360-degree view of the customer by working with different types of data, and build an efficient fraud detection system for a financial institution. You will also develop a system in Hadoop to improve the effectiveness of marketing campaigns. Build a.
|
520 |
8 |
|
|a Churn detection system for a telecom company, develop an Internet of Things (IoT) system to monitor the environment in a factory, and build a data lake - all making use of the concepts and techniques mentioned in this book. The book covers other technologies and frameworks such as Apache Spark, Hive, Sqoop, and more, and how they can be used in conjunction with Hadoop. You will be able to try out the solutions explained in the book and use the knowledge gained to extend them further in your own problem space.
|
590 |
|
|
|a ProQuest Ebook Central
|b Ebook Central Academic Complete
|
630 |
0 |
0 |
|a Apache Hadoop.
|
630 |
0 |
7 |
|a Apache Hadoop
|2 fast
|
650 |
|
0 |
|a Electronic data processing
|x Distributed processing.
|
650 |
|
0 |
|a Big data.
|
650 |
|
6 |
|a Traitement réparti.
|
650 |
|
6 |
|a Données volumineuses.
|
650 |
|
7 |
|a Big data
|2 fast
|
650 |
|
7 |
|a Electronic data processing
|x Distributed processing
|2 fast
|
758 |
|
|
|i has work:
|a Hadoop blueprints (Text)
|1 https://id.oclc.org/worldcat/entity/E39PCFwjgrDpjVhJqBX4gbdpHd
|4 https://id.oclc.org/worldcat/ontology/hasWork
|
856 |
4 |
0 |
|u https://ebookcentral.uam.elogim.com/lib/uam-ebooks/detail.action?docID=4709437
|z Texto completo
|
936 |
|
|
|a BATCHLOAD
|
938 |
|
|
|a Askews and Holts Library Services
|b ASKH
|n AH30656343
|
938 |
|
|
|a ProQuest Ebook Central
|b EBLB
|n EBL4709437
|
938 |
|
|
|a ProQuest MyiLibrary Digital eBook Collection
|b IDEB
|n cis34515001
|
994 |
|
|
|a 92
|b IZTAP
|