Pyspark provides many advantages, but it has a few
as well. A few of them are:-
Since Python is an interpreted language, so pyspark code is relatively
as compared to scala.
Pyspark consumes a lot of
and sometimes it might be challenging if there are a large number of processes running.
Expressing a problem in pyspark is
There are a few
that are only
in scala/java but not available in pyspark.
Since spark is written in scala so using pyspark we cannot fully
the internal functioning of Spark.
apache Kafka use cases