Librería:
GreatBookPrices, Columbia, MD, Estados Unidos de America
Calificación del vendedor: 5 de 5 estrellas
Vendedor de AbeBooks desde 6 de abril de 2009
N° de ref. del artículo 50483160-n
Power Through Big Data at Lightning Speed — With Apache Spark.
In a world overflowing with data, Apache Spark stands out as the go-to engine for fast, distributed processing of massive datasets. This hands-on guide introduces you to the core concepts and real-world use cases of big data analytics using Apache Spark, helping you handle data at scale with ease and efficiency.
Whether you're working with batch jobs, real-time streaming, or machine learning pipelines, this book walks you through the practical steps to build scalable applications for modern data problems — using Spark’s APIs in Python (PySpark), Scala, and Java.
🚀 What You’ll Learn:✅ The architecture of Apache Spark and its components (RDDs, DataFrames, Datasets)
✅ Spark vs. Hadoop: key differences and when to use what
✅ Batch and streaming data processing
✅ Data exploration and transformation with Spark SQL
✅ Using PySpark for hands-on big data analysis
✅ Real-time analytics with Spark Streaming and Kafka
✅ Distributed machine learning with MLlib
✅ Running Spark on Hadoop, YARN, and Kubernetes
✅ Performance tuning, memory optimization, and partitioning strategies
✅ End-to-end project: big data ETL pipeline with real datasets
Data engineers and analysts
Big data and cloud professionals
Software developers expanding into analytics
Students learning scalable data processing
Anyone building real-time or batch big data solutions
Leverage the speed of Apache Spark to unlock insights from massive datasets.
Título: Big Data Processing with Apache Spark: Using...
Editorial: Independently published
Año de publicación: 2025
Encuadernación: Encuadernación de tapa blanda
Condición: New
Librería: California Books, Miami, FL, Estados Unidos de America
Condición: New. Print on Demand. Nº de ref. del artículo: I-9798289301697
Cantidad disponible: Más de 20 disponibles