First, what you need to understand is that PySpark is not a Programming Language. It is just a collaboration of Python with Apache Spark. Technically speaking, PySpark is a collaboration of Python API for powerful data processing engine, Apache Spark. Now, Scala is a Programming Language, and often said that it is faster than Python, almost 10x faster than it, but that does not mean Scala is faster than PySpark. The best way to put it would be to say, that Scala Spark is faster than PySpark. Since Apache Spark is written in Scala, so you would be able better utilize the full potential of Apache Spark.
If you are planning to break into the Big Data domain with knowledge of Python, then check out the PySpark tutorial, and if you wish to get certified in it, then check out the PySpark certification course, from Intellipaat. Watch the following video on PySpark Tutorial for Beginners to help you get started.