Advertisement
Butanium

spark error

Dec 13th, 2021
736
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 21.19 KB | None | 0 0
  1. C:\Users\Clement>spark-shell
  2. Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
  3. Setting default log level to "WARN".
  4. To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
  5. 21/12/13 22:39:29 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  6. 21/12/13 22:39:30 ERROR SparkContext: Error initializing SparkContext.
  7. java.lang.reflect.InvocationTargetException
  8. at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
  9. at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
  10. at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source)
  11. at java.lang.reflect.Constructor.newInstance(Unknown Source)
  12. at org.apache.spark.executor.Executor.addReplClassLoaderIfNeeded(Executor.scala:909)
  13. at org.apache.spark.executor.Executor.<init>(Executor.scala:160)
  14. at org.apache.spark.scheduler.local.LocalEndpoint.<init>(LocalSchedulerBackend.scala:64)
  15. at org.apache.spark.scheduler.local.LocalSchedulerBackend.start(LocalSchedulerBackend.scala:132)
  16. at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:220)
  17. at org.apache.spark.SparkContext.<init>(SparkContext.scala:581)
  18. at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2690)
  19. at org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$2(SparkSession.scala:949)
  20. at scala.Option.getOrElse(Option.scala:189)
  21. at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:943)
  22. at org.apache.spark.repl.Main$.createSparkSession(Main.scala:106)
  23. at $line3.$read$$iw$$iw.<init>(<console>:15)
  24. at $line3.$read$$iw.<init>(<console>:42)
  25. at $line3.$read.<init>(<console>:44)
  26. at $line3.$read$.<init>(<console>:48)
  27. at $line3.$read$.<clinit>(<console>)
  28. at $line3.$eval$.$print$lzycompute(<console>:7)
  29. at $line3.$eval$.$print(<console>:6)
  30. at $line3.$eval.$print(<console>)
  31. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  32. at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
  33. at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
  34. at java.lang.reflect.Method.invoke(Unknown Source)
  35. at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:747)
  36. at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1020)
  37. at scala.tools.nsc.interpreter.IMain.$anonfun$interpret$1(IMain.scala:568)
  38. at scala.reflect.internal.util.ScalaClassLoader.asContext(ScalaClassLoader.scala:36)
  39. at scala.reflect.internal.util.ScalaClassLoader.asContext$(ScalaClassLoader.scala:116)
  40. at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:41)
  41. at scala.tools.nsc.interpreter.IMain.loadAndRunReq$1(IMain.scala:567)
  42. at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:594)
  43. at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:564)
  44. at scala.tools.nsc.interpreter.IMain.$anonfun$quietRun$1(IMain.scala:216)
  45. at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
  46. at scala.tools.nsc.interpreter.IMain.quietRun(IMain.scala:216)
  47. at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$2(SparkILoop.scala:83)
  48. at scala.collection.immutable.List.foreach(List.scala:431)
  49. at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$1(SparkILoop.scala:83)
  50. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  51. at scala.tools.nsc.interpreter.ILoop.savingReplayStack(ILoop.scala:97)
  52. at org.apache.spark.repl.SparkILoop.initializeSpark(SparkILoop.scala:83)
  53. at org.apache.spark.repl.SparkILoop.$anonfun$process$4(SparkILoop.scala:165)
  54. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  55. at scala.tools.nsc.interpreter.ILoop.$anonfun$mumly$1(ILoop.scala:166)
  56. at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
  57. at scala.tools.nsc.interpreter.ILoop.mumly(ILoop.scala:163)
  58. at org.apache.spark.repl.SparkILoop.loopPostInit$1(SparkILoop.scala:153)
  59. at org.apache.spark.repl.SparkILoop.$anonfun$process$10(SparkILoop.scala:221)
  60. at org.apache.spark.repl.SparkILoop.withSuppressedSettings$1(SparkILoop.scala:189)
  61. at org.apache.spark.repl.SparkILoop.startup$1(SparkILoop.scala:201)
  62. at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:236)
  63. at org.apache.spark.repl.Main$.doMain(Main.scala:78)
  64. at org.apache.spark.repl.Main$.main(Main.scala:58)
  65. at org.apache.spark.repl.Main.main(Main.scala)
  66. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  67. at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
  68. at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
  69. at java.lang.reflect.Method.invoke(Unknown Source)
  70. at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
  71. at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:955)
  72. at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
  73. at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
  74. at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
  75. at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1043)
  76. at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1052)
  77. at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
  78. Caused by: java.net.URISyntaxException: Illegal character in path at index 31: spark://10.204.160.180:51834/C:\classes
  79. at java.net.URI$Parser.fail(Unknown Source)
  80. at java.net.URI$Parser.checkChars(Unknown Source)
  81. at java.net.URI$Parser.parseHierarchical(Unknown Source)
  82. at java.net.URI$Parser.parse(Unknown Source)
  83. at java.net.URI.<init>(Unknown Source)
  84. at org.apache.spark.repl.ExecutorClassLoader.<init>(ExecutorClassLoader.scala:57)
  85. ... 70 more
  86. 21/12/13 22:39:30 ERROR Utils: Uncaught exception in thread main
  87. java.lang.NullPointerException
  88. at org.apache.spark.scheduler.local.LocalSchedulerBackend.org$apache$spark$scheduler$local$LocalSchedulerBackend$$stop(LocalSchedulerBackend.scala:173)
  89. at org.apache.spark.scheduler.local.LocalSchedulerBackend.stop(LocalSchedulerBackend.scala:144)
  90. at org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:927)
  91. at org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:2516)
  92. at org.apache.spark.SparkContext.$anonfun$stop$12(SparkContext.scala:2086)
  93. at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1442)
  94. at org.apache.spark.SparkContext.stop(SparkContext.scala:2086)
  95. at org.apache.spark.SparkContext.<init>(SparkContext.scala:677)
  96. at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2690)
  97. at org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$2(SparkSession.scala:949)
  98. at scala.Option.getOrElse(Option.scala:189)
  99. at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:943)
  100. at org.apache.spark.repl.Main$.createSparkSession(Main.scala:106)
  101. at $line3.$read$$iw$$iw.<init>(<console>:15)
  102. at $line3.$read$$iw.<init>(<console>:42)
  103. at $line3.$read.<init>(<console>:44)
  104. at $line3.$read$.<init>(<console>:48)
  105. at $line3.$read$.<clinit>(<console>)
  106. at $line3.$eval$.$print$lzycompute(<console>:7)
  107. at $line3.$eval$.$print(<console>:6)
  108. at $line3.$eval.$print(<console>)
  109. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  110. at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
  111. at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
  112. at java.lang.reflect.Method.invoke(Unknown Source)
  113. at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:747)
  114. at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1020)
  115. at scala.tools.nsc.interpreter.IMain.$anonfun$interpret$1(IMain.scala:568)
  116. at scala.reflect.internal.util.ScalaClassLoader.asContext(ScalaClassLoader.scala:36)
  117. at scala.reflect.internal.util.ScalaClassLoader.asContext$(ScalaClassLoader.scala:116)
  118. at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:41)
  119. at scala.tools.nsc.interpreter.IMain.loadAndRunReq$1(IMain.scala:567)
  120. at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:594)
  121. at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:564)
  122. at scala.tools.nsc.interpreter.IMain.$anonfun$quietRun$1(IMain.scala:216)
  123. at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
  124. at scala.tools.nsc.interpreter.IMain.quietRun(IMain.scala:216)
  125. at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$2(SparkILoop.scala:83)
  126. at scala.collection.immutable.List.foreach(List.scala:431)
  127. at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$1(SparkILoop.scala:83)
  128. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  129. at scala.tools.nsc.interpreter.ILoop.savingReplayStack(ILoop.scala:97)
  130. at org.apache.spark.repl.SparkILoop.initializeSpark(SparkILoop.scala:83)
  131. at org.apache.spark.repl.SparkILoop.$anonfun$process$4(SparkILoop.scala:165)
  132. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  133. at scala.tools.nsc.interpreter.ILoop.$anonfun$mumly$1(ILoop.scala:166)
  134. at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
  135. at scala.tools.nsc.interpreter.ILoop.mumly(ILoop.scala:163)
  136. at org.apache.spark.repl.SparkILoop.loopPostInit$1(SparkILoop.scala:153)
  137. at org.apache.spark.repl.SparkILoop.$anonfun$process$10(SparkILoop.scala:221)
  138. at org.apache.spark.repl.SparkILoop.withSuppressedSettings$1(SparkILoop.scala:189)
  139. at org.apache.spark.repl.SparkILoop.startup$1(SparkILoop.scala:201)
  140. at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:236)
  141. at org.apache.spark.repl.Main$.doMain(Main.scala:78)
  142. at org.apache.spark.repl.Main$.main(Main.scala:58)
  143. at org.apache.spark.repl.Main.main(Main.scala)
  144. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  145. at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
  146. at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
  147. at java.lang.reflect.Method.invoke(Unknown Source)
  148. at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
  149. at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:955)
  150. at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
  151. at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
  152. at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
  153. at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1043)
  154. at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1052)
  155. at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
  156. 21/12/13 22:39:30 WARN MetricsSystem: Stopping a MetricsSystem that is not running
  157. 21/12/13 22:39:30 ERROR Main: Failed to initialize Spark session.
  158. java.lang.reflect.InvocationTargetException
  159. at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
  160. at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
  161. at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source)
  162. at java.lang.reflect.Constructor.newInstance(Unknown Source)
  163. at org.apache.spark.executor.Executor.addReplClassLoaderIfNeeded(Executor.scala:909)
  164. at org.apache.spark.executor.Executor.<init>(Executor.scala:160)
  165. at org.apache.spark.scheduler.local.LocalEndpoint.<init>(LocalSchedulerBackend.scala:64)
  166. at org.apache.spark.scheduler.local.LocalSchedulerBackend.start(LocalSchedulerBackend.scala:132)
  167. at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:220)
  168. at org.apache.spark.SparkContext.<init>(SparkContext.scala:581)
  169. at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2690)
  170. at org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$2(SparkSession.scala:949)
  171. at scala.Option.getOrElse(Option.scala:189)
  172. at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:943)
  173. at org.apache.spark.repl.Main$.createSparkSession(Main.scala:106)
  174. at $line3.$read$$iw$$iw.<init>(<console>:15)
  175. at $line3.$read$$iw.<init>(<console>:42)
  176. at $line3.$read.<init>(<console>:44)
  177. at $line3.$read$.<init>(<console>:48)
  178. at $line3.$read$.<clinit>(<console>)
  179. at $line3.$eval$.$print$lzycompute(<console>:7)
  180. at $line3.$eval$.$print(<console>:6)
  181. at $line3.$eval.$print(<console>)
  182. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  183. at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
  184. at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
  185. at java.lang.reflect.Method.invoke(Unknown Source)
  186. at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:747)
  187. at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1020)
  188. at scala.tools.nsc.interpreter.IMain.$anonfun$interpret$1(IMain.scala:568)
  189. at scala.reflect.internal.util.ScalaClassLoader.asContext(ScalaClassLoader.scala:36)
  190. at scala.reflect.internal.util.ScalaClassLoader.asContext$(ScalaClassLoader.scala:116)
  191. at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:41)
  192. at scala.tools.nsc.interpreter.IMain.loadAndRunReq$1(IMain.scala:567)
  193. at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:594)
  194. at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:564)
  195. at scala.tools.nsc.interpreter.IMain.$anonfun$quietRun$1(IMain.scala:216)
  196. at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
  197. at scala.tools.nsc.interpreter.IMain.quietRun(IMain.scala:216)
  198. at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$2(SparkILoop.scala:83)
  199. at scala.collection.immutable.List.foreach(List.scala:431)
  200. at org.apache.spark.repl.SparkILoop.$anonfun$initializeSpark$1(SparkILoop.scala:83)
  201. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  202. at scala.tools.nsc.interpreter.ILoop.savingReplayStack(ILoop.scala:97)
  203. at org.apache.spark.repl.SparkILoop.initializeSpark(SparkILoop.scala:83)
  204. at org.apache.spark.repl.SparkILoop.$anonfun$process$4(SparkILoop.scala:165)
  205. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  206. at scala.tools.nsc.interpreter.ILoop.$anonfun$mumly$1(ILoop.scala:166)
  207. at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:206)
  208. at scala.tools.nsc.interpreter.ILoop.mumly(ILoop.scala:163)
  209. at org.apache.spark.repl.SparkILoop.loopPostInit$1(SparkILoop.scala:153)
  210. at org.apache.spark.repl.SparkILoop.$anonfun$process$10(SparkILoop.scala:221)
  211. at org.apache.spark.repl.SparkILoop.withSuppressedSettings$1(SparkILoop.scala:189)
  212. at org.apache.spark.repl.SparkILoop.startup$1(SparkILoop.scala:201)
  213. at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:236)
  214. at org.apache.spark.repl.Main$.doMain(Main.scala:78)
  215. at org.apache.spark.repl.Main$.main(Main.scala:58)
  216. at org.apache.spark.repl.Main.main(Main.scala)
  217. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  218. at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
  219. at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
  220. at java.lang.reflect.Method.invoke(Unknown Source)
  221. at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
  222. at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:955)
  223. at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
  224. at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
  225. at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
  226. at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1043)
  227. at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1052)
  228. at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
  229. Caused by: java.net.URISyntaxException: Illegal character in path at index 31: spark://10.204.160.180:51834/C:\classes
  230. at java.net.URI$Parser.fail(Unknown Source)
  231. at java.net.URI$Parser.checkChars(Unknown Source)
  232. at java.net.URI$Parser.parseHierarchical(Unknown Source)
  233. at java.net.URI$Parser.parse(Unknown Source)
  234. at java.net.URI.<init>(Unknown Source)
  235. at org.apache.spark.repl.ExecutorClassLoader.<init>(ExecutorClassLoader.scala:57)
  236. ... 70 more
  237. 21/12/13 22:39:30 ERROR Utils: Uncaught exception in thread shutdown-hook-0
  238. java.lang.ExceptionInInitializerError
  239. at org.apache.spark.executor.Executor.stop(Executor.scala:333)
  240. at org.apache.spark.executor.Executor.$anonfun$stopHookReference$1(Executor.scala:76)
  241. at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)
  242. at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)
  243. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  244. at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)
  245. at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)
  246. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  247. at scala.util.Try$.apply(Try.scala:213)
  248. at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)
  249. at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)
  250. at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
  251. at java.util.concurrent.FutureTask.run(Unknown Source)
  252. at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
  253. at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
  254. at java.lang.Thread.run(Unknown Source)
  255. Caused by: java.lang.NullPointerException
  256. at org.apache.spark.shuffle.ShuffleBlockPusher$.<init>(ShuffleBlockPusher.scala:465)
  257. at org.apache.spark.shuffle.ShuffleBlockPusher$.<clinit>(ShuffleBlockPusher.scala)
  258. ... 16 more
  259. 21/12/13 22:39:30 WARN ShutdownHookManager: ShutdownHook '$anon$2' failed, java.util.concurrent.ExecutionException: java.lang.ExceptionInInitializerError
  260. java.util.concurrent.ExecutionException: java.lang.ExceptionInInitializerError
  261. at java.util.concurrent.FutureTask.report(Unknown Source)
  262. at java.util.concurrent.FutureTask.get(Unknown Source)
  263. at org.apache.hadoop.util.ShutdownHookManager.executeShutdown(ShutdownHookManager.java:124)
  264. at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:95)
  265. Caused by: java.lang.ExceptionInInitializerError
  266. at org.apache.spark.executor.Executor.stop(Executor.scala:333)
  267. at org.apache.spark.executor.Executor.$anonfun$stopHookReference$1(Executor.scala:76)
  268. at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)
  269. at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)
  270. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  271. at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)
  272. at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)
  273. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  274. at scala.util.Try$.apply(Try.scala:213)
  275. at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)
  276. at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)
  277. at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
  278. at java.util.concurrent.FutureTask.run(Unknown Source)
  279. at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
  280. at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
  281. at java.lang.Thread.run(Unknown Source)
  282. Caused by: java.lang.NullPointerException
  283. at org.apache.spark.shuffle.ShuffleBlockPusher$.<init>(ShuffleBlockPusher.scala:465)
  284. at org.apache.spark.shuffle.ShuffleBlockPusher$.<clinit>(ShuffleBlockPusher.scala)
  285. ... 16 more
  286.  
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement