This project is a hands-on collection of notebooks, code snippets, and exercises focused on learning Apache Spark with Python (PySpark). It includes my notes and experiments while exploring core Spark concepts, transformations, actions, DataFrame API, and more.
PySpark is the Python API for Apache Spark, a powerful open-source distributed computing engine used for large-scale data processing and analytics. PySpark allows you to leverage the power of distributed computing using Python.
- β Introduction to Spark & PySpark
- β SparkContext & SparkSession
- β RDDs (Resilient Distributed Datasets)
- β DataFrames & Datasets
- β Transformations vs Actions
- β Reading/Writing: JSON, CSV, Parquet
- β PySpark SQL & Queries
- β GroupBy, Aggregations, Joins
- β Handling Nulls & Missing Data
- β User-Defined Functions (UDFs)
- β Window Functions
- β Data Partitioning & Performance Optimization
- β Intro to MLlib (Optional)
I follow a "Learn by Doing" approach. Each notebook contains:
β Detailed explanations
π§ͺ Hands-on code examples
π Real-world case studies