1

I have upgraded the spark-version to 2.4.3, but I found this exception

Am I missing anything?

    To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
    java.lang.NoClassDefFoundError: org/apache/hadoop/fs/StorageStatistics
      at java.lang.Class.forName0(Native Method)
      at java.lang.Class.forName(Class.java:348)
      at org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.java:2134)
      at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:2099)
      at org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2193)
      at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2654)
      at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2667)
      at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94)
      at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703)
      at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685)
      at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373)
      at org.apache.spark.util.Utils$.getHadoopFileSystem(Utils.scala:1866)
      at org.apache.spark.scheduler.EventLoggingListener.<init>(EventLoggingListener.scala:71)
      at org.apache.spark.SparkContext.<init>(SparkContext.scala:521)
      at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2520)
      at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:935)
      at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:926)
      at scala.Option.getOrElse(Option.scala:121)
      at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:926)
      at org.apache.spark.repl.Main$.createSparkSession(Main.scala:106)
      ... 62 elided
    Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.fs.StorageStatistics
      at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
      at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
      at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:349)
      at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
      ... 82 more
    <console>:14: error: not found: value spark
           import spark.implicits._
                  ^
    <console>:14: error: not found: value spark
           import spark.sql
    These are the versions for aws sdk and hadoop-aws
    I have upgraded the spark-version to 2.4.3, but I found this exception 

Am I missing anything?

aws sdk jars that I have configured

hadoop-aws-3.2.0.jar
aws-java-sdk-bundle-1.11.619.jar
Andrei Tanana
  • 7,932
  • 1
  • 27
  • 36
  • Please check this, root cause is same https://stackoverflow.com/questions/44411493/java-lang-noclassdeffounderror-org-apache-hadoop-fs-storagestatistics – maogautam Aug 28 '19 at 21:01

1 Answers1

0

I got similar error, after adding hadoop-common, the issue is gone. So in my case like this:

  • hadoop-common
  • hadoop-aws
  • aws-java-sdk-bundle
Hongbo Miao
  • 45,290
  • 60
  • 174
  • 267