A DataFrame can be defined as a dataset designed as named columns,i.e. is a distributed collection of data. Conceptually, it is equivalent to relational tables.
Spark functionality contains some core parts and CSV is one of them.
A DataFrame may be created from a variety of input sources including the CSV text files, JSON files, etc.
To load a CSV file as a DataFrame write these command on your Spark shell :
df=spark.read.format("csv").option("header","true").load("/home/amit/uo.csv")
You can refer the following video if you want more information regarding the same: