I am trying to use Sparkmeausre to check the performance of my Pyspark code. I am using Pycharm Community edition on windows 10, with Pyspark properly configured. I did "pip install sparkmeasure" and sparkmeasure was sucessfully installed. Now when I am trying to run this snippet of code.
from pyspark import SparkConf , SparkContext
from pyspark.context import SparkContext
from pyspark.sql.session import SparkSession
from sparkmeasure import StageMetrics
sc = SparkContext(master = "local" , appName = "sparkdemo")
spark = SparkSession(sc)
sm = StageMetrics(spark)
I am getting the error.
File "C:/Users/nj123/PycharmProjects/pythonProject/sparkdemo.py", line 9, in <module>
sm = StageMetrics(spark)
File "C:\Users\nj123\PycharmProjects\pythonProject\venv\lib\site-
packages\sparkmeasure\stagemetrics.py", line 15, in __init__
self.stagemetrics = self.sc._jvm.ch.cern.sparkmeasure.StageMetrics(self.sparksession._jsparkSession)
TypeError: 'JavaPackage' object is not callable
How to resolve this error and to configure sparkmeasure to Pycharm correctly?