Spark and RDD User Handbook
Are you a programmer experimenting with in-memory computation on large clusters? If yes, then you must take Spark as well as RDD into your consideration. This Spark and RDD cheat sheet are designed for the one who has already started learning about memory management and using Spark as a tool. This sheet will be a handy reference for them.
You can also download the printable PDF of this Spark & RDD cheat sheet
Now, don’t worry if you are a beginner and have no idea about how Spark and RDD work. This cheat sheet includes all concepts you must know, from the basics, and will give you a quick reference to all of them.
Key Concepts
- Apache Spark: It is an open-source, Hadoop-compatible, fast and expressive cluster computing platform.
- Resilient Distributed Datasets (RDDs): The core concept in Apache Spark is RDDs, which are the immutable distributed collections of data that are partitioned across machines in a cluster.
- Transformation: It is an operation performed on an RDD, such as filter(), map(), or union(), which yields another RDD.
- Action: It is an operation that triggers a computation such as count(), first(), take(n), or collect().
- Partition: It is a logical division of data stored on a node in a cluster.
RDD Components
- SparkContext: It holds a connection with Spark Cluster Management.
- Driver: The process of running the main() function of an application and creating the SparkContext is managed by the driver.
- Worker: Any node which can run the program on the cluster is called a worker.
- Cluster Manager: A cluster manager allocates resources to each application in a driver program. There are three types of cluster managers supported by Apache Spark:
Shared Variables in Spark
exampleAccumulator = sparkContext.accumulator(1)
exampleAccumulator.add(5)
Learn Apache Spark from Big Data and Spark Online Course in Hyderabad and be an Apache Spark Specialist!
Watch this Spark Video for Beginners:
Unified Libraries in Spark
- Spark SQL: It is a Spark module that allows working with structured data. Data querying is supported by SQL or HQL.
- Spark Streaming: It is used to build a scalable application that provides fault-tolerant streaming. It also processes using web server logs, Facebook logs, etc. in real-time.
- MLlib (Machine Learning): It is a scalable Machine Learning library and provides various algorithms for classification, regression, clustering, etc.
- GraphX: It is an API for graphs. This module can efficiently find the shortest path for static graphs.
Components of Spark
- Executors: Executors comprise multiple tasks; basically, it is a JVM process sitting on all nodes. Executors receive the tasks, deserialize them, and run them. Executors utilize cache so that the tasks can be run faster.
- Tasks: Jars, along with the code, are referred to as tasks.
- Nodes: Nodes consist of multiple executors.
- RDDs: It is a big data structure that is used to represent data, which cannot be stored on a single machine. Hence, the data is distributed, partitioned, and split across multiple computers.
- Inputs: Every RDD is made up of some inputs such as a text file, Hadoop file, etc.
- Output: The output of a function in Spark can produce an RDD; it is functional since a function, one after the other, receives an input RDD and outputs an output RDD.
Want to grasp detailed knowledge of Hadoop? Read this extensive Spark Tutorial!
Get 100% Hike!
Master Most in Demand Skills Now!
Commonly Used Transformations
Function |
Description |
map(function) |
Returns a new RDD by applying the function on each data element |
filter(function) |
Returns a new dataset formed by selecting those elements of the source on which the function returns true |
filterByRange(lower, upper) |
Returns an RDD with elements in the specified range, upper to lower |
flatMap(function) |
Similar to the map function but returns a sequence, instead of a value |
reduceByKey(function,[num Tasks]) |
Aggregates the values of a key using a function |
groupByKey([num Tasks]) |
Converts (K,V) to (K, <iterable V>) |
distinct([num Tasks]) |
Eliminates duplicates from an RDD |
mapPartitions(function) |
Similar to map but runs separately on each partition of an RDD |
mapPartitionsWithIndex(function) |
Similar to the map partition but also provides the function with an integer value representing the index of the partition |
sample(withReplacement, fraction, seed) |
Samples a fraction of data using the given random number generating seeds |
union() |
Returns a new RDD containing all elements and arguments of the source RDD |
intersection() |
Returns a new RDD that contains an intersection of elements in the datasets |
Cartesian() |
Returns the Cartesian product of all pairs of elements |
subtract() |
Returns a new RDD created by removing the elements from the source RDD with common arguments |
join(RDD,[numTasks]) |
Joins two elements of the dataset with common arguments; when invoked on (A,B) and (A,C), it creates a new RDD, (A,(B,C)) |
cogroup(RDD,[numTasks]) |
Converts (A,B) to (A, <iterable B>) |
Commonly Used Actions
Function |
Description |
count() |
Gets the number of data elements in an RDD |
collect() |
Gets all data elements of an RDD as an array |
reduce(function) |
Aggregates data elements into an RDD by taking two arguments and returning one |
take(n) |
Fetches the first n elements of an RDD |
foreach(function) |
Executes the function for each data element of an RDD |
first() |
Retrieves the first data element of an RDD |
saveastextfile(path) |
Writes the content of an RDD to a text file, or a set of text files, in the local system |
takeordered(n, [ordering]) |
Returns the first n elements of an RDD using either the natural order or a custom comparator |
Prepare yourself with these Apache Spark Interview Questions and Answers and excel in your career!
Persistence Methods
Function |
Description |
cache() |
Avoids unnecessary recomputation; it is similar to persist(MEMORY_ONLY) |
persist([Storage Level]) |
Persists an RDD with the default storage level |
unpersist() |
Marks an RDD as non-persistent and removes the block from memory and disk |
checkpoint() |
Saves a file inside the checkpoint directory and removes all the references of its parent RDD |
RDD Persistence Methods
Function |
Description |
MEMORY_ONLY (default level) |
Stores an RDD in an available cluster memory as a deserialized Java object |
MEMORY_AND_DISK |
Stores an RDD as a deserialized Java object; if the RDD does not fit in the cluster memory, it stores the partitions on the disk and reads them |
MEMORY_ONLY_SER |
Stores an RDD as a serialized Java object; it is more CPU intensive |
MEMORY_ONLY_DISK_SER |
Similar to the above but stores in a disk when the memory is not sufficient |
DISC_ONLY |
Stores an RDD only on the disk |
MEMORY_ONLY_2, MEMORY_AND_
DISK_2, etc. |
Similar to other levels, except that partitions are replicated on two slave nodes |
With this, you have come to the end of the Spark and RDD Cheat Sheet. To get in-depth knowledge, check out our interactive, online Apache Spark Training that comes with 24/7 support to guide you throughout your learning period. Intellipaat’s Apache Spark training includes Spark Streaming, Spark SQL, Spark RDDs, and Spark Machine Learning libraries (Spark MLlib).