Intellipaat Back

Explore Courses Blog Tutorials Interview Questions
0 votes
2 views
in Big Data Hadoop & Spark by (11.4k points)

We are reading data from MongoDB Collection. Collection column has two different values (e.g.: (bson.Int64,int) (int,float) ).

I am trying to get a datatype using pyspark.

My problem is some columns have different datatype.

Assume quantity and weight are the columns

quantity           weight
---------          --------
12300              656
123566000000       789.6767
1238               56.22
345                23
345566677777789    21


Actually we didn't defined data type for any column of mongo collection.

1 Answer

0 votes
by (32.3k points)

In order to get the data types of your DataFrame columns, you can use dtypes i.e :

>>> df.dtypes

[('age', 'int'), ('name', 'string')]

It shows that your column age is of type int and name is of type string.

Related questions

31k questions

32.8k answers

501 comments

693 users

Browse Categories

...