Rdd

Engineering

Spark Scala - RDD Operations

Resilient Distributed Datasets (RDDs) are Spark’s original abstraction for distributed data processing. While DataFrames and Datasets have become the preferred API for most workloads, understanding…

Read more →
Python

PySpark - RDD Broadcast Variables

Broadcast variables provide an efficient mechanism for sharing read-only data across all nodes in a Spark cluster. Without broadcasting, Spark serializes and sends data with each task, creating…

Read more →
Python

PySpark - RDD join Operations

• RDD joins in PySpark support multiple join types (inner, outer, left outer, right outer) through operations on PairRDDs, where data must be structured as key-value tuples before joining

Read more →
Python

PySpark - Pair RDD Operations

• Pair RDDs are the foundation for distributed key-value operations in PySpark, enabling efficient aggregations, joins, and grouping across partitions through hash-based data distribution.

Read more →
Python

PySpark - Create RDD from Text File

Resilient Distributed Datasets (RDDs) represent PySpark’s fundamental abstraction for distributed data processing. While DataFrames have become the preferred API for structured data, RDDs remain…

Read more →
Python

PySpark - Convert RDD to DataFrame

RDDs (Resilient Distributed Datasets) represent Spark’s low-level API, offering fine-grained control over distributed data. DataFrames build on RDDs while adding schema information and query…

Read more →
Python

PySpark - Create DataFrame from RDD

• DataFrames provide significant performance advantages over RDDs through Catalyst optimizer and Tungsten execution engine, making conversion worthwhile for complex transformations and SQL operations.

Read more →