Back

Explore Courses Blog Tutorials Interview Questions
0 votes
2 views
in Big Data Hadoop & Spark by (11.4k points)

We are reading data from MongoDB Collection. Collection column has two different values (e.g.: (bson.Int64,int) (int,float) ).

I am trying to get a datatype using pyspark.

My problem is some columns have different datatype.

Assume quantity and weight are the columns

quantity           weight
---------          --------
12300              656
123566000000       789.6767
1238               56.22
345                23
345566677777789    21


Actually we didn't defined data type for any column of mongo collection.

1 Answer

0 votes
by (32.3k points)

In order to get the data types of your DataFrame columns, you can use dtypes i.e :

>>> df.dtypes

[('age', 'int'), ('name', 'string')]

It shows that your column age is of type int and name is of type string.

Related questions

Welcome to Intellipaat Community. Get your technical queries answered by top developers!

30.5k questions

32.5k answers

500 comments

108k users

Browse Categories

...