Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- C:\Users\Clement>spark-shell
- Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
- Setting default log level to "WARN".
- To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
- 21/12/13 22:39:29 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
- 21/12/13 22:39:30 ERROR SparkContext: Error initializing SparkContext.
- java.lang.reflect.InvocationTargetException
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source)
- at java.lang.reflect.Constructor.newInstance(Unknown Source)
- at org.apache.spark.executor.Executor.addReplClassLoaderIfNeeded(Executor.scala:909)
- at org.apache.spark.executor.Executor.<init>(Executor.scala:160)
- at org.apache.spark.scheduler.local.LocalEndpoint.<init>(LocalSchedulerBackend.scala:64)
- at org.apache.spark.scheduler.local.LocalSchedulerBackend.start(LocalSchedulerBackend.scala:132)
- at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:220)
- at org.apache.spark.SparkContext.<init>(SparkContext.scala:581)
- at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2690)
- at org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$2(SparkSession.scala:949)
- at scala.Option.getOrElse(Option.scala:189)
- at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:943)
- at org.apache.spark.repl.Main$.createSparkSession(Main.scala:106)
- at $line3.$read$$iw$$iw.<init>(<console>:15)
- at $line3.$read$$iw.<init>(<console>:42)
- at $line3.$read.<init>(<console>:44)
- at $line3.$read$.<init>(<console>:48)
- at $line3.$read$.<clinit>(<console>)
- at $line3.$eval$.$print$lzycompute(<console>:7)
- at $line3.$eval$.$print(<console>:6)
- at $line3.$eval.$print(<console>)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
- at java.lang.reflect.Method.invoke(Unknown Source)
- at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:747)
- at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1020)
- at scala.tools.nsc.interpreter.IMain.$anonfun$interpret$1(IMain.scala:568)
- at scala.reflect.internal.util.ScalaClassLoader.asContext(ScalaClassLoader.scala:36)
- at scala.reflect.internal.util.ScalaClassLoader.asContext$(ScalaClassLoader.scala:116)
- at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:41)
- at scala.tools.nsc.interpreter.IMain.loadAndRunReq$1(IMain.scala:567)
- at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:594)
- at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:564)
- at scala.tools.nsc.interpreter.IMain.$anonfun$quietRun$1(IMain.scala:216)
- at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
- at scala.tools.nsc.interpreter.IMain.quietRun(IMain.scala:216)
- at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$2(SparkILoop.scala:83)
- at scala.collection.immutable.List.foreach(List.scala:431)
- at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$1(SparkILoop.scala:83)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at scala.tools.nsc.interpreter.ILoop.savingReplayStack(ILoop.scala:97)
- at org.apache.spark.repl.SparkILoop.initializeSpark(SparkILoop.scala:83)
- at org.apache.spark.repl.SparkILoop.$anonfun$process$4(SparkILoop.scala:165)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at scala.tools.nsc.interpreter.ILoop.$anonfun$mumly$1(ILoop.scala:166)
- at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
- at scala.tools.nsc.interpreter.ILoop.mumly(ILoop.scala:163)
- at org.apache.spark.repl.SparkILoop.loopPostInit$1(SparkILoop.scala:153)
- at org.apache.spark.repl.SparkILoop.$anonfun$process$10(SparkILoop.scala:221)
- at org.apache.spark.repl.SparkILoop.withSuppressedSettings$1(SparkILoop.scala:189)
- at org.apache.spark.repl.SparkILoop.startup$1(SparkILoop.scala:201)
- at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:236)
- at org.apache.spark.repl.Main$.doMain(Main.scala:78)
- at org.apache.spark.repl.Main$.main(Main.scala:58)
- at org.apache.spark.repl.Main.main(Main.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
- at java.lang.reflect.Method.invoke(Unknown Source)
- at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
- at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:955)
- at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
- at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
- at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
- at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1043)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1052)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- Caused by: java.net.URISyntaxException: Illegal character in path at index 31: spark://10.204.160.180:51834/C:\classes
- at java.net.URI$Parser.fail(Unknown Source)
- at java.net.URI$Parser.checkChars(Unknown Source)
- at java.net.URI$Parser.parseHierarchical(Unknown Source)
- at java.net.URI$Parser.parse(Unknown Source)
- at java.net.URI.<init>(Unknown Source)
- at org.apache.spark.repl.ExecutorClassLoader.<init>(ExecutorClassLoader.scala:57)
- ... 70 more
- 21/12/13 22:39:30 ERROR Utils: Uncaught exception in thread main
- java.lang.NullPointerException
- at org.apache.spark.scheduler.local.LocalSchedulerBackend.org$apache$spark$scheduler$local$LocalSchedulerBackend$$stop(LocalSchedulerBackend.scala:173)
- at org.apache.spark.scheduler.local.LocalSchedulerBackend.stop(LocalSchedulerBackend.scala:144)
- at org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:927)
- at org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:2516)
- at org.apache.spark.SparkContext.$anonfun$stop$12(SparkContext.scala:2086)
- at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1442)
- at org.apache.spark.SparkContext.stop(SparkContext.scala:2086)
- at org.apache.spark.SparkContext.<init>(SparkContext.scala:677)
- at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2690)
- at org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$2(SparkSession.scala:949)
- at scala.Option.getOrElse(Option.scala:189)
- at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:943)
- at org.apache.spark.repl.Main$.createSparkSession(Main.scala:106)
- at $line3.$read$$iw$$iw.<init>(<console>:15)
- at $line3.$read$$iw.<init>(<console>:42)
- at $line3.$read.<init>(<console>:44)
- at $line3.$read$.<init>(<console>:48)
- at $line3.$read$.<clinit>(<console>)
- at $line3.$eval$.$print$lzycompute(<console>:7)
- at $line3.$eval$.$print(<console>:6)
- at $line3.$eval.$print(<console>)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
- at java.lang.reflect.Method.invoke(Unknown Source)
- at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:747)
- at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1020)
- at scala.tools.nsc.interpreter.IMain.$anonfun$interpret$1(IMain.scala:568)
- at scala.reflect.internal.util.ScalaClassLoader.asContext(ScalaClassLoader.scala:36)
- at scala.reflect.internal.util.ScalaClassLoader.asContext$(ScalaClassLoader.scala:116)
- at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:41)
- at scala.tools.nsc.interpreter.IMain.loadAndRunReq$1(IMain.scala:567)
- at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:594)
- at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:564)
- at scala.tools.nsc.interpreter.IMain.$anonfun$quietRun$1(IMain.scala:216)
- at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
- at scala.tools.nsc.interpreter.IMain.quietRun(IMain.scala:216)
- at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$2(SparkILoop.scala:83)
- at scala.collection.immutable.List.foreach(List.scala:431)
- at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$1(SparkILoop.scala:83)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at scala.tools.nsc.interpreter.ILoop.savingReplayStack(ILoop.scala:97)
- at org.apache.spark.repl.SparkILoop.initializeSpark(SparkILoop.scala:83)
- at org.apache.spark.repl.SparkILoop.$anonfun$process$4(SparkILoop.scala:165)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at scala.tools.nsc.interpreter.ILoop.$anonfun$mumly$1(ILoop.scala:166)
- at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
- at scala.tools.nsc.interpreter.ILoop.mumly(ILoop.scala:163)
- at org.apache.spark.repl.SparkILoop.loopPostInit$1(SparkILoop.scala:153)
- at org.apache.spark.repl.SparkILoop.$anonfun$process$10(SparkILoop.scala:221)
- at org.apache.spark.repl.SparkILoop.withSuppressedSettings$1(SparkILoop.scala:189)
- at org.apache.spark.repl.SparkILoop.startup$1(SparkILoop.scala:201)
- at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:236)
- at org.apache.spark.repl.Main$.doMain(Main.scala:78)
- at org.apache.spark.repl.Main$.main(Main.scala:58)
- at org.apache.spark.repl.Main.main(Main.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
- at java.lang.reflect.Method.invoke(Unknown Source)
- at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
- at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:955)
- at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
- at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
- at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
- at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1043)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1052)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- 21/12/13 22:39:30 WARN MetricsSystem: Stopping a MetricsSystem that is not running
- 21/12/13 22:39:30 ERROR Main: Failed to initialize Spark session.
- java.lang.reflect.InvocationTargetException
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source)
- at java.lang.reflect.Constructor.newInstance(Unknown Source)
- at org.apache.spark.executor.Executor.addReplClassLoaderIfNeeded(Executor.scala:909)
- at org.apache.spark.executor.Executor.<init>(Executor.scala:160)
- at org.apache.spark.scheduler.local.LocalEndpoint.<init>(LocalSchedulerBackend.scala:64)
- at org.apache.spark.scheduler.local.LocalSchedulerBackend.start(LocalSchedulerBackend.scala:132)
- at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:220)
- at org.apache.spark.SparkContext.<init>(SparkContext.scala:581)
- at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2690)
- at org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$2(SparkSession.scala:949)
- at scala.Option.getOrElse(Option.scala:189)
- at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:943)
- at org.apache.spark.repl.Main$.createSparkSession(Main.scala:106)
- at $line3.$read$$iw$$iw.<init>(<console>:15)
- at $line3.$read$$iw.<init>(<console>:42)
- at $line3.$read.<init>(<console>:44)
- at $line3.$read$.<init>(<console>:48)
- at $line3.$read$.<clinit>(<console>)
- at $line3.$eval$.$print$lzycompute(<console>:7)
- at $line3.$eval$.$print(<console>:6)
- at $line3.$eval.$print(<console>)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
- at java.lang.reflect.Method.invoke(Unknown Source)
- at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:747)
- at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1020)
- at scala.tools.nsc.interpreter.IMain.$anonfun$interpret$1(IMain.scala:568)
- at scala.reflect.internal.util.ScalaClassLoader.asContext(ScalaClassLoader.scala:36)
- at scala.reflect.internal.util.ScalaClassLoader.asContext$(ScalaClassLoader.scala:116)
- at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:41)
- at scala.tools.nsc.interpreter.IMain.loadAndRunReq$1(IMain.scala:567)
- at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:594)
- at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:564)
- at scala.tools.nsc.interpreter.IMain.$anonfun$quietRun$1(IMain.scala:216)
- at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
- at scala.tools.nsc.interpreter.IMain.quietRun(IMain.scala:216)
- at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$2(SparkILoop.scala:83)
- at scala.collection.immutable.List.foreach(List.scala:431)
- at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$1(SparkILoop.scala:83)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at scala.tools.nsc.interpreter.ILoop.savingReplayStack(ILoop.scala:97)
- at org.apache.spark.repl.SparkILoop.initializeSpark(SparkILoop.scala:83)
- at org.apache.spark.repl.SparkILoop.$anonfun$process$4(SparkILoop.scala:165)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at scala.tools.nsc.interpreter.ILoop.$anonfun$mumly$1(ILoop.scala:166)
- at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
- at scala.tools.nsc.interpreter.ILoop.mumly(ILoop.scala:163)
- at org.apache.spark.repl.SparkILoop.loopPostInit$1(SparkILoop.scala:153)
- at org.apache.spark.repl.SparkILoop.$anonfun$process$10(SparkILoop.scala:221)
- at org.apache.spark.repl.SparkILoop.withSuppressedSettings$1(SparkILoop.scala:189)
- at org.apache.spark.repl.SparkILoop.startup$1(SparkILoop.scala:201)
- at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:236)
- at org.apache.spark.repl.Main$.doMain(Main.scala:78)
- at org.apache.spark.repl.Main$.main(Main.scala:58)
- at org.apache.spark.repl.Main.main(Main.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
- at java.lang.reflect.Method.invoke(Unknown Source)
- at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
- at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:955)
- at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
- at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
- at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
- at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1043)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1052)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- Caused by: java.net.URISyntaxException: Illegal character in path at index 31: spark://10.204.160.180:51834/C:\classes
- at java.net.URI$Parser.fail(Unknown Source)
- at java.net.URI$Parser.checkChars(Unknown Source)
- at java.net.URI$Parser.parseHierarchical(Unknown Source)
- at java.net.URI$Parser.parse(Unknown Source)
- at java.net.URI.<init>(Unknown Source)
- at org.apache.spark.repl.ExecutorClassLoader.<init>(ExecutorClassLoader.scala:57)
- ... 70 more
- 21/12/13 22:39:30 ERROR Utils: Uncaught exception in thread shutdown-hook-0
- java.lang.ExceptionInInitializerError
- at org.apache.spark.executor.Executor.stop(Executor.scala:333)
- at org.apache.spark.executor.Executor.$anonfun$stopHookReference$1(Executor.scala:76)
- at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)
- at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)
- at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at scala.util.Try$.apply(Try.scala:213)
- at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)
- at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)
- at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
- at java.util.concurrent.FutureTask.run(Unknown Source)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
- at java.lang.Thread.run(Unknown Source)
- Caused by: java.lang.NullPointerException
- at org.apache.spark.shuffle.ShuffleBlockPusher$.<init>(ShuffleBlockPusher.scala:465)
- at org.apache.spark.shuffle.ShuffleBlockPusher$.<clinit>(ShuffleBlockPusher.scala)
- ... 16 more
- 21/12/13 22:39:30 WARN ShutdownHookManager: ShutdownHook '$anon$2' failed, java.util.concurrent.ExecutionException: java.lang.ExceptionInInitializerError
- java.util.concurrent.ExecutionException: java.lang.ExceptionInInitializerError
- at java.util.concurrent.FutureTask.report(Unknown Source)
- at java.util.concurrent.FutureTask.get(Unknown Source)
- at org.apache.hadoop.util.ShutdownHookManager.executeShutdown(ShutdownHookManager.java:124)
- at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:95)
- Caused by: java.lang.ExceptionInInitializerError
- at org.apache.spark.executor.Executor.stop(Executor.scala:333)
- at org.apache.spark.executor.Executor.$anonfun$stopHookReference$1(Executor.scala:76)
- at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)
- at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)
- at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at scala.util.Try$.apply(Try.scala:213)
- at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)
- at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)
- at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
- at java.util.concurrent.FutureTask.run(Unknown Source)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
- at java.lang.Thread.run(Unknown Source)
- Caused by: java.lang.NullPointerException
- at org.apache.spark.shuffle.ShuffleBlockPusher$.<init>(ShuffleBlockPusher.scala:465)
- at org.apache.spark.shuffle.ShuffleBlockPusher$.<clinit>(ShuffleBlockPusher.scala)
- ... 16 more
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement