Search
Carol McDonald

Datasets, DataFrames, and Spark SQL for Processing of Tabular Data

August 19, 2020

Original Post Information:

"authorDisplayName": "Carol McDonald",
"publish": "2018-10-24T07:00:00.000Z",
"tags": "spark"

With Apache Spark 2.0 and later versions, big improvements were implemented to make Spark easier to program and execute faster. The Spark SQL and the Dataset/DataFrame APIs provide ease of use, space efficiency, and performance gains with Spark SQL's optimized execution engine. In this blog post we will give an introduction to Spark Datasets, DataFrames and Spark SQL. This is part 2 of a multi-blog series. You can read part 1 here.

Editor’s Note: MapR products referenced are now part of the HPE Ezmeral Data Fabric.

A Spark Dataset is a distributed collection of typed objects, which are partitioned across multiple nodes in a cluster and can be operated on in parallel. Datasets can be created from MapR XD files, MapR Database tables, or MapR Event Store topics, and can be cached, allowing reuse across parallel operations. A Dataset can be manipulated using functional transformations (map, flatMap, filter, etc.) and/or Spark SQL. A DataFrame is a Dataset of Row objects and represents a table of data with rows and columns. A DataFrame consists of partitions, each of which is a range of rows in cache on a data node.

The SparkSession Object

As discussed before, a Spark application runs as independent processes, coordinated by the SparkSession object in the driver program. The entry point to programming in Spark is the org.apache.spark.sql.SparkSession class, which you use to create a SparkSession object as shown below:

val spark =
SparkSession.builder().appName("example").master("local[*]".getOrCreate()

If you are using the spark-shell or a notebook, the SparkSession object is already created and available as the variable Spark.

Interactive Analysis with the Spark Shell

The Spark shell provides an easy way to learn Spark interactively. You can start the shell with the following command:

$ /[installation path]/bin/spark-shell --master local[2]

You can enter the code from the rest of this chapter into the Spark shell. Outputs from the shell are prefaced with ---result---.

Exploring U.S. Flight Data with Spark Datasets and DataFrames

To go over some core concepts of Spark Datasets, we will be using some flight information from the United States Department of Transportation. Later, we will use this same data to predict flight delays, so we want to explore the flight attributes that contribute most to flight delays. Using Spark Datasets, we will explore the data to answer questions like: which airline carriers, days of the week, originating airport, and hours of the day have the highest number of flight delays, when a delay is greater than 40 minutes?

The flight data is in JSON files, with each flight having the following information:

  • id: ID composed of carrier, date, origin, destination, flight number
  • dofW: day of week (1=Monday, 7=Sunday)        
  • carrier: carrier code        
  • origin: origin airport code        
  • dest: destination airport code
  • crsdephour: scheduled departure hour        
  • crsdeptime: scheduled departure time        
  • depdelay: departure delay in minutes        
  • crsarrtime: scheduled arrival time        
  • arrdelay: arrival delay minutes        
  • crselapsedtime: elapsed time        
  • dist: distance        

It appears in the following format:

{    
"_id": "AA_2017-01-01_ATL_LGA_1678",
"dofW": 7,
"carrier": "AA",
"origin": "ATL",
"dest": "LGA",
"crsdephour": 17,
"crsdeptime": 1700,
"depdelay": 0.0,
"crsarrtime": 1912,
"arrdelay": 0.0,
"crselapsedtime": 132.0,
"dist": 762.0
}

(The complete data and code for all examples are available here.)

Loading Data from a File into a Dataset

With the SparkSession read method, we can read data from a file into a DataFrame, specifying the file type, file path, and input options for the schema. The schema can optionally be inferred from the contents of the JSON file, but you will get better performance and accuracy by specifying the schema.

import org.apache.spark.sql.types._
import org.apache.spark.sql._
import org.apache.spark.sql.functions._

val schema = StructType(Array(
    StructField("_id", StringType, true),
    StructField("dofW", IntegerType, true),
    StructField("carrier", StringType, true),
    StructField("origin", StringType, true),
    StructField("dest", StringType, true),
    StructField("crsdephour", IntegerType, true),
    StructField("crsdeptime", DoubleType, true),
    StructField("crsarrtime", DoubleType, true),
    StructField("crselapsedtime", DoubleType, true),
    StructField("label", DoubleType, true),
    StructField("pred_dtree", DoubleType, true)
  ))
var file = "maprfs:///data/flights.json"

val df = spark.read.format("json").option("inferSchema", "false").schema(schema).load(file)

---result:---
df: org.apache.spark.sql.DataFrame = [_id: string, dofW: int ... 10 more fields]

The take method returns an array with objects from this Dataset, which we see is of type Row.

df.take(1)

---result:---
Array[org.apache.spark.sql.Row] =
Array([ATL_LGA_2017-01-01_17_AA_1678, 7, AA, ATL, LGA, 17, 1700.0, 0.0, 1912.0, 0.0, 132.0, 762.0])

If we supply a case class with the as method when loading the data, then the data is read into a Dataset of typed objects corresponding to the case class.

case class Flight(_id: String, dofW: Integer, carrier: String, origin: String, dest: String, crsdephour: Integer, crsdeptime: Double, depdelay: Double,crsarrtime: Double, arrdelay: Double, crselapsedtime: Double, dist: Double) extends Serializable

val df = spark.read.format("json").option("inferSchema", "false").schema(schema).load(file).as[Flight]

---result---:
df: org.apache.spark.sql.Dataset[Flight] = [_id: string, dofW: int ... 10 more fields]

Now the take method returns an array of Flight objects.

df.take(1)

---result:---
Array[Flight] = Array(Flight(ATL_LGA_2017-01-01_17_AA_1678, 7,AA,ATL,LGA,17,1700.0,0.0,1912.0,0.0,132.0,762.0))

Transformations and Actions

There are two types of operations you can perform on a Dataset:

  • transformations: create a new Dataset from the current Dataset
  • actions: trigger computation and return a result to the driver program

Transformations are lazily evaluated, which means they are not computed immediately. A transformation is executed only when it is triggered by an action. Once an action has run and the value is returned, the Dataset is no longer in memory, unless you call the cache method on the Dataset. If you will reuse a Dataset for more than one action, you should cache it.

Datasets and Type Safety

Datasets are composed of typed objects, which means that transformation syntax errors (like a typo in the method name) and analysis errors (like an incorrect input variable type) can be caught at compile time. DataFrames are composed of untyped Row objects, which means that only syntax errors can be caught at compile time. Spark SQL is composed of a string, which means that syntax errors and analysis errors are only caught at runtime. Datasets save a developer’s time by catching errors sooner, even while typing when using an IDE.

Dataset Transformations

Here is a list of some commonly used typed transformations, which can be used on Datasets of typed objects (Dataset[T]).

Transformation
Description
mapReturns new Dataset with result of applying input function to each element
filterReturns new Dataset containing elements where input function is true
groupByKeyReturns a KeyValueGroupedDataset where the data is grouped by the given key function

This example filter transformation on the flight Dataset returns a Dataset with flights that departed at 10 AM. The take action returns an array of flight objects to the driver program.

df.filter(flight => flight.crsdephour == 10).take(3)

---result:---
Array[Flight] = Array(Flight(ORD_DEN_2017-01-01_AA_2300, 7,AA,ORD,DEN,10,1005.0,5.0,1145.0,3.0,160.0,888.0), Flight(MIA_ORD_2017-01-01_AA_2439,7,AA,MIA,ORD,10, 1005.0,4.0,1231.0,0.0,206.0,1197.0))

DataFrame Transformations

Here is a list of some commonly used untyped transformations, which can be used on Dataframes (Dataset[Row]).

Transformation
Description
selectSelects a set of columns
joinJoin with another DataFrame, using the given join expression
groupByGroups the DataFrame, using the specified columns

This groupBy transformation example groups the flight Dataset by carrier, then the count action counts the number of flights for each carrier. The show action prints out the resulting DataFrame rows in tabular format.

df.groupBy("carrier").count().show()

---result:---
+-------+-----+
|carrier|count|
+-------+-----+
|     UA|18873|
|     AA|10031|
|     DL|10055|
|     WN| 2389|
+-------+-----+

Here is a list of some commonly used Dataset actions.

Action
Description
show(n)Displays the first n rows in a tabular form
take(n)Returns the first n objects in the Dataset in an array
countReturns the number of rows in the Dataset

Here is an example using typed and untyped transformations and actions to get the destinations with the highest number of departure delays, where a delay is greater than 40 minutes. We count the departure delays greater than 40 minutes by destination and sort them with the highest first.

df.filter($"depdelay" > 40).groupBy("dest").count()
.orderBy(desc("count")).show(3)

---result:---
+----+-----+
|dest|count|
+----+-----+
| SFO|  711|
| EWR|  620|
| ORD|  593|
+----+-----+

Exploring the Flight Dataset with Spark SQL

Now let’s explore the flight Dataset using Spark SQL and DataFrame transformations. After we register the DataFrame as a SQL temporary view, we can use SQL functions on the SparkSession to run SQL queries, which will return the results as a DataFrame. We cache the DataFrame, since we will reuse it and because Spark can cache DataFrames or Tables in columnar format in memory, which can improve memory usage and performance.

// cache DataFrame in columnar format in memory
df.cache

// create Table view of DataFrame for Spark SQL
df.createOrReplaceTempView("flights")

// cache flights table in columnar format in memory
spark.catalog.cacheTable("flights")

Below, we display information for the top five longest departure delays with Spark SQL and with DataFrame transformations (where a delay is considered greater than 40 minutes):

// Spark SQL
spark.sql("select carrier,origin, dest, depdelay,crsdephour, dist, dofW from flights where depdelay > 40 order by depdelay desc limit 5").show

// same query using DataFrame transformations

df.select($"carrier",$"origin",$"dest",$"depdelay", $"crsdephour").filter($"depdelay" > 40).orderBy(desc( "depdelay" )).show(5)

---result:---
+-------+------+----+--------+----------+
|carrier|origin|dest|depdelay|crsdephour|
+-------+------+----+--------+----------+
|     AA|   SFO| ORD|  1440.0|         8|
|     DL|   BOS| ATL|  1185.0|        17|
|     UA|   DEN| EWR|  1138.0|        12|
|     DL|   ORD| ATL|  1087.0|        19|
|     UA|   MIA| EWR|  1072.0|        20|
+-------+------+----+--------+----------+

Below, we display the average departure delay by carrier:

// DataFrame transformations

df.groupBy("carrier").agg(avg("depdelay")).show

---result:---
+-------+------------------+
|carrier|     avg(depdelay)|
+-------+------------------+
|     UA|17.477878450696764|
|     AA| 10.45768118831622|
|     DL|15.316061660865241|
|     WN|13.491000418585182|
+-------+------------------+

With a notebook like Zeppelin or Jupyter, you can also display the SQL results in graph formats.

// Spark SQL
%sql select carrier, avg(depdelay)
 from flights
 group by carrier

Let’s explore this data for flight delays, when the departure delay is greater than 40 minutes. Below, we see that United Airlines and Delta have the highest count of flight delays for January and February 2017 (the training set).

// __Count of Departure Delays by Carrier (where delay=40 minutes)__

df.filter($"depdelay" > 40)
.groupBy("carrier").count.orderBy(desc( "count")).show(5)

---result:---
+-------+-----+
|carrier|count|
+-------+-----+
|     UA| 2420|
|     DL| 1043|
|     AA|  757|
|     WN|  244|
+-------+-----+
// Count of Departure Delays by Carrier (where delay=40 minutes)

%sql
select carrier, count(depdelay)
from flights where depdelay > 40
group by carrier

In the query below, we see that Monday (1), Tuesday (2), and Sunday (7) have the highest count of flight delays.

// Count of Departure Delays by Day of the Week

%sql
select dofW, count(depdelay)
from flights where depdelay > 40
group by dofW

In the query below, we see that the hours between 13:00-19:00 have the highest count of flight delays.

%sql
select crsdephour, count(depdelay)
from flights where depdelay > 40
group by crsdephour order by crsdephour

In the query below, we see that the originating airports, Chicago and Atlanta, have the highest count of flight delays.

%sql
select origin, count(depdelay)
from flights where depdelay > 40
group by origin
ORDER BY count(depdelay) desc

In the query below, we see the count of departure delays by origin and destination. The routes ORD->SFO and DEN->SFO have the highest delays, maybe because of weather in January and February. Adding weather to this Dataset would give better results.

%sql
select origin, dest, count(depdelay)
from flights where depdelay > 40
group by origin, dest
ORDER BY count(depdelay) desc

Summary

You have now learned how to load data into Spark Datasets and DataFrames and how to explore tabular data with Spark SQL. These code examples can be reused as the foundation to solve many types of business problems.

Related

Ted Dunning & Ellen Friedman

3 ways a data fabric enables a data-first approach

Mar 15, 2022
Nicolas Perez

A Functional Approach to Logging in Apache Spark

Feb 5, 2021
Cenz Wong

Getting Started with DataTaps in Kubernetes Pods

Jul 6, 2021
Kiran Kumar Mavatoor

Accessing HPE Ezmeral Data Fabric Object Storage from Spring Boot S3 Micro Service deployed in K3s cluster

Sep 13, 2021
Carol McDonald

An Inside Look at the Components of a Recommendation Engine

Jan 22, 2021
Carol McDonald

Analyzing Flight Delays with Apache Spark GraphFrames and MapR Database

Dec 16, 2020
Nicolas Perez

Apache Spark as a Distributed SQL Engine

Jan 7, 2021
Carol McDonald

Apache Spark Machine Learning Tutorial

Nov 25, 2020

HPE Developer Newsletter

Stay in the loop.

Sign up for the HPE Developer Newsletter or visit the Newsletter Archive to see past content.

By clicking on “Subscribe Now”, I agree to HPE sending me personalized email communication about HPE and select HPE-Partner products, services, offers and events. I understand that my email address will be used in accordance with HPE Privacy Statement. You may unsubscribe from receiving HPE and HPE-Partner news and offers at any time by clicking on the Unsubscribe button at the bottom of the newsletter.

For more information on how HPE manages, uses, and protects your personal data please refer to HPE Privacy Statement.