0

I create a spark cluster environment and i am facing a freezing issue when i try to show a dataframe,

builder = SparkSession.builder \
   .appName('Contabilidade > Conta Cosif') \
   .config("spark.jars", "/home/dir/op-cast-lramos/.ivy2/jars/io.delta_delta-core_2.12-2.2.0.jar,../drivers/zstd-jni-1.5.2-1.jar") \
   .config("spark.sql.extensions", "io.delta.sql.DeltaSparkSessionExtension") \
   .config("spark.sql.catalog.spark_catalog", "org.apache.spark.sql.delta.catalog.DeltaCatalog")\
   .config("spark.sql.debug.maxToStringFields", 1000)\
   .master('spark://server:7077')

spark = configure_spark_with_delta_pip(builder).getOrCreate()
data = spark.range(0, 5)
data.write.format("delta").save("/datalake/workspace/storage/dag002")
df= spark.read.format("delta").load("/datalake/workspace/storage/dag002")
    
df.show()  ==> in this part of code , i am facing the freezing...   

My environment:

  • Red Hat Linux 4.18.0-425.3.1.el8.x86_64
  • Python 3.7.11
      ____              __
     / __/__  ___ _____/ /__
    _\ \/ _ \/ _ `/ __/  '_/
   /___/ .__/\_,_/_/ /_/\_\   version 3.3.1
      /_/

Using Scala version 2.12.15, OpenJDK 64-Bit Server VM, 12
  • delta lake 2.12-2.2.0
Alex Ott
  • 80,552
  • 8
  • 87
  • 132

0 Answers0