|
|
|
|
LEADER |
00000cam a22000007a 4500 |
001 |
OR_on1398279353 |
003 |
OCoLC |
005 |
20231017213018.0 |
006 |
m o d |
007 |
cr |n||||||||| |
008 |
230921s2023 enk o 001 0 eng d |
040 |
|
|
|a YDX
|b eng
|c YDX
|d OCLKB
|d EBLCP
|d OCLCO
|d OCLKB
|d OCLKB
|d ORMDA
|
019 |
|
|
|a 1398243786
|
020 |
|
|
|a 9781801076418
|q (electronic bk.)
|
020 |
|
|
|a 1801076413
|q (electronic bk.)
|
020 |
|
|
|z 1801070490
|
020 |
|
|
|z 9781801070492
|
035 |
|
|
|a (OCoLC)1398279353
|z (OCoLC)1398243786
|
037 |
|
|
|a 9781801070492
|b O'Reilly Media
|
050 |
|
4 |
|a QA76.73.P98
|
082 |
0 |
4 |
|a 005.13/3
|2 23/eng/20231010
|
049 |
|
|
|a UAMI
|
100 |
1 |
|
|a Lipp, Brian,
|e author.
|
245 |
1 |
0 |
|a MODERN DATA ARCHITECTURES WITH PYTHON
|h [electronic resource] :
|b a practical guide to building and deploying data pipelines, data warehouses, and data lakes with Python /
|c Brian Lipp.
|
250 |
|
|
|a 1st edition.
|
260 |
|
|
|a Birmingham, UK :
|b Packt Publishing Ltd.,
|c 2023.
|
300 |
|
|
|a 1 online resource
|
336 |
|
|
|a text
|b txt
|2 rdacontent
|
337 |
|
|
|a computer
|b c
|2 rdamedia
|
338 |
|
|
|a online resource
|b cr
|2 rdacarrier
|
500 |
|
|
|a Includes index.
|
505 |
0 |
|
|a Cover -- Title Page -- Copyright and Credits -- Dedications -- Contributors -- Table of Contents -- Preface -- Part 1: Fundamental Data Knowledge -- Chapter 1: Modern Data Processing Architecture -- Technical requirements -- Databases, data warehouses, and data lakes -- OLTP -- OLAP -- Data lakes -- Event stores -- File formats -- Data platform architecture at a high level -- Comparing the Lambda and Kappa architectures -- Lambda architecture -- Kappa architecture -- Lakehouse and Delta architectures -- Lakehouses -- The seven central tenets
|
505 |
8 |
|
|a The medallion data pattern and the Delta architecture -- Data mesh theory and practice -- Defining terms -- The four principles of data mesh -- Summary -- Practical lab -- Solution -- Chapter 2: Understanding Data Analytics -- Technical requirements -- Setting up your environment -- Python -- venv -- Graphviz -- Workflow initialization -- Cleaning and preparing your data -- Duplicate values -- Working with nulls -- Using RegEx -- Outlier identification -- Casting columns -- Fixing column names -- Complex data types -- Data documentation -- diagrams -- Data lineage graphs -- Data modeling patterns
|
505 |
8 |
|
|a Relational -- Dimensional modeling -- Key terms -- OBT -- Practical lab -- Loading the problem data -- Solution -- Summary -- Part 2: Data Engineering Toolset -- Chapter 3: Apache Spark Deep Dive -- Technical requirements -- Setting up your environment -- Python, AWS, and Databricks -- Databricks CLI -- Cloud data storage -- Object storage -- Relational -- NoSQL -- Spark architecture -- Introduction to Apache Spark -- Key components -- Working with partitions -- Shuffling partitions -- Caching -- Broadcasting -- Job creation pipeline -- Delta Lake -- Transaction log
|
505 |
8 |
|
|a Grouping tables with databases -- Table -- Adding speed with Z-ordering -- Bloom filters -- Practical lab -- Problem 1 -- Problem 2 -- Problem 3 -- Solution -- Summary -- Chapter 4: Batch and Stream Data Processing Using PySpark -- Technical requirements -- Setting up your environment -- Python, AWS, and Databricks -- Databricks CLI -- Batch processing -- Partitioning -- Data skew -- Reading data -- Spark schemas -- Making decisions -- Removing unwanted columns -- Working with data in groups -- The UDF -- Stream processing -- Reading from disk -- Debugging -- Writing to disk
|
505 |
8 |
|
|a Batch stream hybrid -- Delta streaming -- Batch processing in a stream -- Practical lab -- Setup -- Creating fake data -- Problem 1 -- Problem 2 -- Problem 3 -- Solution -- Solution 1 -- Solution 2 -- Solution 3 -- Summary -- Chapter 5: Streaming Data with Kafka -- Technical requirements -- Setting up your environment -- Python, AWS, and Databricks -- Databricks CLI -- Confluent Kafka -- Signing up -- Kafka architecture -- Topics -- Partitions -- Brokers -- Producers -- Consumers -- Schema Registry -- Kafka Connect -- Spark and Kafka -- Practical lab -- Solution -- Summary
|
520 |
|
|
|a Build scalable and reliable data ecosystems using Data Mesh, Databricks Spark, and Kafka Key Features Develop modern data skills used in emerging technologies Learn pragmatic design methodologies such as Data Mesh and data lakehouses Gain a deeper understanding of data governance Purchase of the print or Kindle book includes a free PDF eBook Book Description Modern Data Architectures with Python will teach you how to seamlessly incorporate your machine learning and data science work streams into your open data platforms. You'll learn how to take your data and create open lakehouses that work with any technology using tried-and-true techniques, including the medallion architecture and Delta Lake. Starting with the fundamentals, this book will help you build pipelines on Databricks, an open data platform, using SQL and Python. You'll gain an understanding of notebooks and applications written in Python using standard software engineering tools such as git, pre-commit, Jenkins, and Github. Next, you'll delve into streaming and batch-based data processing using Apache Spark and Confluent Kafka. As you advance, you'll learn how to deploy your resources using infrastructure as code and how to automate your workflows and code development. Since any data platform's ability to handle and work with AI and ML is a vital component, you'll also explore the basics of ML and how to work with modern MLOps tooling. Finally, you'll get hands-on experience with Apache Spark, one of the key data technologies in today's market. By the end of this book, you'll have amassed a wealth of practical and theoretical knowledge to build, manage, orchestrate, and architect your data ecosystems. What you will learn Understand data patterns including delta architecture Discover how to increase performance with Spark internals Find out how to design critical data diagrams Explore MLOps with tools such as AutoML and MLflow Get to grips with building data products in a data mesh Discover data governance and build confidence in your data Introduce data visualizations and dashboards into your data practice Who this book is for This book is for developers, analytics engineers, and managers looking to further develop a data ecosystem within their organization. While they're not prerequisites, basic knowledge of Python and prior experience with data will help you to read and follow along with the examples.
|
590 |
|
|
|a O'Reilly
|b O'Reilly Online Learning: Academic/Public Library Edition
|
650 |
|
0 |
|a Python (Computer program language)
|
650 |
|
0 |
|a Data structures (Computer science)
|
650 |
|
0 |
|a Big data.
|
776 |
0 |
8 |
|i Print version:
|z 1801070490
|z 9781801070492
|w (OCoLC)1396693127
|
856 |
4 |
0 |
|u https://learning.oreilly.com/library/view/~/9781801070492/?ar
|z Texto completo (Requiere registro previo con correo institucional)
|
938 |
|
|
|a YBP Library Services
|b YANK
|n 20497915
|
938 |
|
|
|b OCKB
|z pqebk.perpetual,1caf4064-7041-41d0-8c33-fd828b7909bc-emi
|
994 |
|
|
|a 92
|b IZTAP
|