Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- --- command line
- Windows PowerShell
- Copyright (C) Microsoft Corporation. All rights reserved.
- Install the latest PowerShell for new features and improvements! https://aka.ms/PSWindows
- PS C:\Users\Lenovo\Desktop\BD_Elastic_Leyla_Baxridinova> docker compose up
- [+] Running 35/3
- ✔ elasticsearch Pulled 3354.9s
- ✔ logstash Pulled 3288.6s
- ✔ kibana Pulled 2166.2s
- [+] Running 1/1
- ✘ Network bd_elastic_leyla_baxridinova_bigdata_network Error 15.0s
- failed to create network bd_elastic_leyla_baxridinova_bigdata_network: Error response from daemon: plugin "local" not found
- PS C:\Users\Lenovo\Desktop\BD_Elastic_Leyla_Baxridinova> docker compose up
- [+] Running 4/4
- ✔ Network bd_elastic_leyla_baxridinova_default Created 0.0s
- ✔ Container es Created 0.1s
- ✔ Container kib Created 0.1s
- ✔ Container log Created 0.1s
- Attaching to es, kib, log
- log | Using bundled JDK: /usr/share/logstash/jdk
- log | OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
- Gracefully stopping... (press Ctrl+C again to force)
- [+] Stopping 3/3
- ✔ Container log Stopped 0.7s
- ✔ Container kib Stopped 0.6s
- ✔ Container es Stopped 0.4s
- PS C:\Users\Lenovo\Desktop\BD_Elastic_Leyla_Baxridinova> docker ps -a
- CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
- 2f581b737b07 kibana:7.16.1 "/bin/tini -- /usr/l…" 40 seconds ago Up 16 seconds 0.0.0.0:5601->5601/tcp kib
- 264e1808c009 logstash:7.16.1 "/usr/local/bin/dock…" 40 seconds ago Up 16 seconds 0.0.0.0:5000->5000/tcp, 0.0.0.0:5044->5044/tcp, 0.0.0.0:9600->9600/tcp, 0.0.0.0:5000->5000/udp log
- 6d87bd20cd3e elasticsearch:7.16.1 "/bin/tini -- /usr/l…" 40 seconds ago Up 16 seconds (health: starting) 0.0.0.0:9200->9200/tcp, 0.0.0.0:9300->9300/tcp es
- PS C:\Users\Lenovo\Desktop\BD_Elastic_Leyla_Baxridinova> docker inspect bd_elastic_leyla_baxridinova_default
- [
- {
- "Name": "bd_elastic_leyla_baxridinova_default",
- "Id": "dd4b26d8eef6e45f91a8690b9e1dee3f46911d083f7aa30c3436f6d9d04fddd7",
- "Created": "2024-05-19T07:17:41.837439845Z",
- "Scope": "local",
- "Driver": "bridge",
- "EnableIPv6": false,
- "IPAM": {
- "Driver": "default",
- "Options": null,
- "Config": [
- {
- "Subnet": "172.19.0.0/16",
- "Gateway": "172.19.0.1"
- }
- ]
- },
- "Internal": false,
- "Attachable": false,
- "Ingress": false,
- "ConfigFrom": {
- "Network": ""
- },
- "ConfigOnly": false,
- "Containers": {
- "2f581b737b073fbf446b4425a23f76442fa66ca77ae9210229709b149930406e": {
- "Name": "kib",
- "EndpointID": "396c1875dfd567fd1ab3d2036a15a49612b9a62f9260f4d6a8a3511d7977fc1e",
- "MacAddress": "02:42:ac:13:00:03",
- "IPv4Address": "172.19.0.3/16",
- "IPv6Address": ""
- },
- "6d87bd20cd3e3f86e67d863e264f57818bbd3151aba0ee6eaec62ad0136a9b04": {
- "Name": "es",
- "EndpointID": "59ba7b8221eb3a5c91b41b448725deb80a1d1e39a4e1a563a47e0bbec6b1ade2",
- "MacAddress": "02:42:ac:13:00:02",
- "IPv4Address": "172.19.0.2/16",
- "IPv6Address": ""
- }
- },
- "Options": {},
- "Labels": {
- "com.docker.compose.network": "default",
- "com.docker.compose.project": "bd_elastic_leyla_baxridinova",
- "com.docker.compose.version": "2.27.0"
- }
- }
- ]
- PS C:\Users\Lenovo\Desktop\BD_Elastic_Leyla_Baxridinova>
- --- stacktrace
- Connected to the target VM, address: '127.0.0.1:55141', transport: 'socket'
- Connecting to Elasticsearch...
- WARNING: An illegal reflective access operation has occurred
- WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform (file:/C:/Users/Lenovo/.m2/repository/org/apache/spark/spark-unsafe_2.12/3.1.2/spark-unsafe_2.12-3.1.2.jar) to constructor java.nio.DirectByteBuffer(long,int)
- WARNING: Please consider reporting this to the maintainers of org.apache.spark.unsafe.Platform
- WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
- WARNING: All illegal access operations will be denied in a future release
- Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
- 24/05/19 13:21:32 INFO SparkContext: Running Spark version 3.1.2
- 24/05/19 13:21:32 INFO ResourceUtils: ==============================================================
- 24/05/19 13:21:32 INFO ResourceUtils: No custom resources configured for spark.driver.
- 24/05/19 13:21:32 INFO ResourceUtils: ==============================================================
- 24/05/19 13:21:32 INFO SparkContext: Submitted application: StreamingElastic
- 24/05/19 13:21:32 INFO ResourceProfile: Default ResourceProfile created, executor resources: Map(cores -> name: cores, amount: 1, script: , vendor: , memory -> name: memory, amount: 1024, script: , vendor: , offHeap -> name: offHeap, amount: 0, script: , vendor: ), task resources: Map(cpus -> name: cpus, amount: 1.0)
- 24/05/19 13:21:32 INFO ResourceProfile: Limiting resource is cpu
- 24/05/19 13:21:32 INFO ResourceProfileManager: Added ResourceProfile id: 0
- 24/05/19 13:21:32 INFO SecurityManager: Changing view acls to: Lenovo
- 24/05/19 13:21:32 INFO SecurityManager: Changing modify acls to: Lenovo
- 24/05/19 13:21:32 INFO SecurityManager: Changing view acls groups to:
- 24/05/19 13:21:32 INFO SecurityManager: Changing modify acls groups to:
- 24/05/19 13:21:32 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(Lenovo); groups with view permissions: Set(); users with modify permissions: Set(Lenovo); groups with modify permissions: Set()
- 24/05/19 13:21:33 INFO Utils: Successfully started service 'sparkDriver' on port 55181.
- 24/05/19 13:21:33 INFO SparkEnv: Registering MapOutputTracker
- 24/05/19 13:21:33 INFO SparkEnv: Registering BlockManagerMaster
- 24/05/19 13:21:33 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
- 24/05/19 13:21:33 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
- 24/05/19 13:21:33 INFO SparkEnv: Registering BlockManagerMasterHeartbeat
- 24/05/19 13:21:33 INFO DiskBlockManager: Created local directory at C:\Users\Lenovo\AppData\Local\Temp\blockmgr-7c0bf709-1376-42b9-86c9-eaf06a20e829
- 24/05/19 13:21:33 INFO MemoryStore: MemoryStore started with capacity 4.6 GiB
- 24/05/19 13:21:33 INFO SparkEnv: Registering OutputCommitCoordinator
- 24/05/19 13:21:33 INFO Utils: Successfully started service 'SparkUI' on port 4040.
- 24/05/19 13:21:33 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at http://Lenovo:4040
- 24/05/19 13:21:34 INFO Executor: Starting executor ID driver on host Lenovo
- 24/05/19 13:21:34 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 55232.
- 24/05/19 13:21:34 INFO NettyBlockTransferService: Server created on Lenovo:55232
- 24/05/19 13:21:34 INFO BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
- 24/05/19 13:21:34 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(driver, Lenovo, 55232, None)
- 24/05/19 13:21:34 INFO BlockManagerMasterEndpoint: Registering block manager Lenovo:55232 with 4.6 GiB RAM, BlockManagerId(driver, Lenovo, 55232, None)
- 24/05/19 13:21:34 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(driver, Lenovo, 55232, None)
- 24/05/19 13:21:34 INFO BlockManager: Initialized BlockManager: BlockManagerId(driver, Lenovo, 55232, None)
- Preparing simple data...
- 24/05/19 13:21:34 INFO SharedState: Setting hive.metastore.warehouse.dir ('null') to the value of spark.sql.warehouse.dir ('file:/D:/IdeaProjects/University/Year%202/Semester%204/Big%20data/learning_bigdata_elasticsearch_spark_to_es_simple_app-main/spark-warehouse').
- 24/05/19 13:21:34 INFO SharedState: Warehouse path is 'file:/D:/IdeaProjects/University/Year%202/Semester%204/Big%20data/learning_bigdata_elasticsearch_spark_to_es_simple_app-main/spark-warehouse'.
- 24/05/19 13:21:35 INFO Version: Elasticsearch Hadoop v8.12.0 [e138d23add]
- 24/05/19 13:21:35 INFO InMemoryFileIndex: It took 46 ms to list leaf files for 1 paths.
- 24/05/19 13:21:35 INFO InMemoryFileIndex: It took 1 ms to list leaf files for 1 paths.
- 24/05/19 13:21:36 INFO FileSourceStrategy: Pushed Filters:
- 24/05/19 13:21:36 INFO FileSourceStrategy: Post-Scan Filters: (length(trim(value#0, None)) > 0)
- 24/05/19 13:21:36 INFO FileSourceStrategy: Output Data Schema: struct<value: string>
- 24/05/19 13:21:37 INFO CodeGenerator: Code generated in 104.9 ms
- 24/05/19 13:21:37 INFO MemoryStore: Block broadcast_0 stored as values in memory (estimated size 174.5 KiB, free 4.6 GiB)
- 24/05/19 13:21:37 INFO MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 27.7 KiB, free 4.6 GiB)
- 24/05/19 13:21:37 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory on Lenovo:55232 (size: 27.7 KiB, free: 4.6 GiB)
- 24/05/19 13:21:37 INFO SparkContext: Created broadcast 0 from csv at SparkJavaElasticStreamTest.java:23
- 24/05/19 13:21:37 INFO FileSourceScanExec: Planning scan with bin packing, max size: 4194304 bytes, open cost is considered as scanning 4194304 bytes.
- 24/05/19 13:21:37 INFO SparkContext: Starting job: csv at SparkJavaElasticStreamTest.java:23
- 24/05/19 13:21:37 INFO DAGScheduler: Got job 0 (csv at SparkJavaElasticStreamTest.java:23) with 1 output partitions
- 24/05/19 13:21:37 INFO DAGScheduler: Final stage: ResultStage 0 (csv at SparkJavaElasticStreamTest.java:23)
- 24/05/19 13:21:37 INFO DAGScheduler: Parents of final stage: List()
- 24/05/19 13:21:37 INFO DAGScheduler: Missing parents: List()
- 24/05/19 13:21:37 INFO DAGScheduler: Submitting ResultStage 0 (MapPartitionsRDD[3] at csv at SparkJavaElasticStreamTest.java:23), which has no missing parents
- 24/05/19 13:21:37 INFO MemoryStore: Block broadcast_1 stored as values in memory (estimated size 10.8 KiB, free 4.6 GiB)
- 24/05/19 13:21:37 INFO MemoryStore: Block broadcast_1_piece0 stored as bytes in memory (estimated size 5.4 KiB, free 4.6 GiB)
- 24/05/19 13:21:37 INFO BlockManagerInfo: Added broadcast_1_piece0 in memory on Lenovo:55232 (size: 5.4 KiB, free: 4.6 GiB)
- 24/05/19 13:21:37 INFO SparkContext: Created broadcast 1 from broadcast at DAGScheduler.scala:1388
- 24/05/19 13:21:37 INFO DAGScheduler: Submitting 1 missing tasks from ResultStage 0 (MapPartitionsRDD[3] at csv at SparkJavaElasticStreamTest.java:23) (first 15 tasks are for partitions Vector(0))
- 24/05/19 13:21:37 INFO TaskSchedulerImpl: Adding task set 0.0 with 1 tasks resource profile 0
- 24/05/19 13:21:37 INFO TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0) (Lenovo, executor driver, partition 0, PROCESS_LOCAL, 4984 bytes) taskResourceAssignments Map()
- 24/05/19 13:21:37 INFO Executor: Running task 0.0 in stage 0.0 (TID 0)
- 24/05/19 13:21:37 INFO FileScanRDD: Reading File path: file:///D:/IdeaProjects/University/Year%202/Semester%204/Big%20data/learning_bigdata_elasticsearch_spark_to_es_simple_app-main/src/main/resources/test/test.csv, range: 0-248, partition values: [empty row]
- 24/05/19 13:21:37 INFO CodeGenerator: Code generated in 11.7618 ms
- 24/05/19 13:21:37 INFO Executor: Finished task 0.0 in stage 0.0 (TID 0). 1659 bytes result sent to driver
- 24/05/19 13:21:37 INFO TaskSetManager: Finished task 0.0 in stage 0.0 (TID 0) in 188 ms on Lenovo (executor driver) (1/1)
- 24/05/19 13:21:37 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose tasks have all completed, from pool
- 24/05/19 13:21:37 INFO DAGScheduler: ResultStage 0 (csv at SparkJavaElasticStreamTest.java:23) finished in 0.298 s
- 24/05/19 13:21:37 INFO DAGScheduler: Job 0 is finished. Cancelling potential speculative or zombie tasks for this job
- 24/05/19 13:21:37 INFO TaskSchedulerImpl: Killing all running tasks in stage 0: Stage finished
- 24/05/19 13:21:37 INFO DAGScheduler: Job 0 finished: csv at SparkJavaElasticStreamTest.java:23, took 0.339042 s
- 24/05/19 13:21:37 INFO CodeGenerator: Code generated in 7.8712 ms
- 24/05/19 13:21:37 INFO FileSourceStrategy: Pushed Filters:
- 24/05/19 13:21:37 INFO FileSourceStrategy: Post-Scan Filters:
- 24/05/19 13:21:37 INFO FileSourceStrategy: Output Data Schema: struct<value: string>
- 24/05/19 13:21:37 INFO MemoryStore: Block broadcast_2 stored as values in memory (estimated size 174.5 KiB, free 4.6 GiB)
- 24/05/19 13:21:37 INFO MemoryStore: Block broadcast_2_piece0 stored as bytes in memory (estimated size 27.7 KiB, free 4.6 GiB)
- 24/05/19 13:21:37 INFO BlockManagerInfo: Added broadcast_2_piece0 in memory on Lenovo:55232 (size: 27.7 KiB, free: 4.6 GiB)
- 24/05/19 13:21:37 INFO SparkContext: Created broadcast 2 from csv at SparkJavaElasticStreamTest.java:23
- 24/05/19 13:21:37 INFO FileSourceScanExec: Planning scan with bin packing, max size: 4194304 bytes, open cost is considered as scanning 4194304 bytes.
- 24/05/19 13:21:37 INFO InMemoryFileIndex: It took 2 ms to list leaf files for 1 paths.
- Writing simple data...
- 24/05/19 13:21:38 INFO FileSourceStrategy: Pushed Filters:
- 24/05/19 13:21:38 INFO FileSourceStrategy: Post-Scan Filters:
- 24/05/19 13:21:38 INFO FileSourceStrategy: Output Data Schema: struct<id: string, franchise_id: string, franchise_name: string, restaurant_franchise_id: string, country: string ... 10 more fields>
- 24/05/19 13:21:38 INFO MemoryStore: Block broadcast_3 stored as values in memory (estimated size 174.4 KiB, free 4.6 GiB)
- 24/05/19 13:21:38 INFO MemoryStore: Block broadcast_3_piece0 stored as bytes in memory (estimated size 27.6 KiB, free 4.6 GiB)
- 24/05/19 13:21:38 INFO BlockManagerInfo: Added broadcast_3_piece0 in memory on Lenovo:55232 (size: 27.6 KiB, free: 4.6 GiB)
- 24/05/19 13:21:38 INFO SparkContext: Created broadcast 3 from rdd at EsSparkSQL.scala:103
- 24/05/19 13:21:38 INFO FileSourceScanExec: Planning scan with bin packing, max size: 4194304 bytes, open cost is considered as scanning 4194304 bytes.
- 24/05/19 13:21:38 INFO SparkContext: Starting job: runJob at EsSparkSQL.scala:103
- 24/05/19 13:21:38 INFO DAGScheduler: Got job 1 (runJob at EsSparkSQL.scala:103) with 1 output partitions
- 24/05/19 13:21:38 INFO DAGScheduler: Final stage: ResultStage 1 (runJob at EsSparkSQL.scala:103)
- 24/05/19 13:21:38 INFO DAGScheduler: Parents of final stage: List()
- 24/05/19 13:21:38 INFO DAGScheduler: Missing parents: List()
- 24/05/19 13:21:38 INFO DAGScheduler: Submitting ResultStage 1 (MapPartitionsRDD[14] at rdd at EsSparkSQL.scala:103), which has no missing parents
- 24/05/19 13:21:38 INFO MemoryStore: Block broadcast_4 stored as values in memory (estimated size 16.2 KiB, free 4.6 GiB)
- 24/05/19 13:21:38 INFO MemoryStore: Block broadcast_4_piece0 stored as bytes in memory (estimated size 7.9 KiB, free 4.6 GiB)
- 24/05/19 13:21:38 INFO BlockManagerInfo: Added broadcast_4_piece0 in memory on Lenovo:55232 (size: 7.9 KiB, free: 4.6 GiB)
- 24/05/19 13:21:38 INFO SparkContext: Created broadcast 4 from broadcast at DAGScheduler.scala:1388
- 24/05/19 13:21:38 INFO DAGScheduler: Submitting 1 missing tasks from ResultStage 1 (MapPartitionsRDD[14] at rdd at EsSparkSQL.scala:103) (first 15 tasks are for partitions Vector(0))
- 24/05/19 13:21:38 INFO TaskSchedulerImpl: Adding task set 1.0 with 1 tasks resource profile 0
- 24/05/19 13:21:38 INFO TaskSetManager: Starting task 0.0 in stage 1.0 (TID 1) (Lenovo, executor driver, partition 0, PROCESS_LOCAL, 4984 bytes) taskResourceAssignments Map()
- 24/05/19 13:21:38 INFO Executor: Running task 0.0 in stage 1.0 (TID 1)
- 24/05/19 13:21:38 INFO CodeGenerator: Code generated in 19.5392 ms
- 24/05/19 13:21:38 WARN Resource: Detected type name in resource [receipt_restaurants/data]. Type names are deprecated and will be removed in a later release.
- 24/05/19 13:21:38 INFO EsDataFrameWriter: Writing to [receipt_restaurants/data]
- 24/05/19 13:21:48 WARN ProcfsMetricsGetter: Exception when trying to compute pagesize, as a result reporting of ProcessTree metrics is stopped
- 24/05/19 13:21:59 INFO HttpMethodDirector: I/O exception (java.net.ConnectException) caught when processing request: Connection timed out: connect
- 24/05/19 13:21:59 INFO HttpMethodDirector: Retrying request
- 24/05/19 13:22:20 INFO HttpMethodDirector: I/O exception (java.net.ConnectException) caught when processing request: Connection timed out: connect
- 24/05/19 13:22:20 INFO HttpMethodDirector: Retrying request
- 24/05/19 13:22:41 INFO HttpMethodDirector: I/O exception (java.net.ConnectException) caught when processing request: Connection timed out: connect
- 24/05/19 13:22:41 INFO HttpMethodDirector: Retrying request
- 24/05/19 13:23:02 ERROR NetworkClient: Node [172.19.0.2:9200] failed (java.net.ConnectException: Connection timed out: connect); no other nodes left - aborting...
- 24/05/19 13:23:02 ERROR Executor: Exception in task 0.0 in stage 1.0 (TID 1)
- org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException: Connection error (check network and/or proxy settings)- all nodes failed; tried [[172.19.0.2:9200]]
- at org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:160)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:442)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:438)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:398)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:402)
- at org.elasticsearch.hadoop.rest.RestClient.get(RestClient.java:178)
- at org.elasticsearch.hadoop.rest.request.GetAliasesRequestBuilder.execute(GetAliasesRequestBuilder.java:68)
- at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:620)
- at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:71)
- at org.elasticsearch.spark.sql.EsSparkSQL$.$anonfun$saveToEs$1(EsSparkSQL.scala:103)
- at org.elasticsearch.spark.sql.EsSparkSQL$.$anonfun$saveToEs$1$adapted(EsSparkSQL.scala:103)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
- at org.apache.spark.scheduler.Task.run(Task.scala:131)
- at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
- at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
- at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
- at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
- at java.base/java.lang.Thread.run(Thread.java:829)
- 24/05/19 13:23:02 WARN TaskSetManager: Lost task 0.0 in stage 1.0 (TID 1) (Lenovo executor driver): org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException: Connection error (check network and/or proxy settings)- all nodes failed; tried [[172.19.0.2:9200]]
- at org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:160)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:442)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:438)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:398)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:402)
- at org.elasticsearch.hadoop.rest.RestClient.get(RestClient.java:178)
- at org.elasticsearch.hadoop.rest.request.GetAliasesRequestBuilder.execute(GetAliasesRequestBuilder.java:68)
- at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:620)
- at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:71)
- at org.elasticsearch.spark.sql.EsSparkSQL$.$anonfun$saveToEs$1(EsSparkSQL.scala:103)
- at org.elasticsearch.spark.sql.EsSparkSQL$.$anonfun$saveToEs$1$adapted(EsSparkSQL.scala:103)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
- at org.apache.spark.scheduler.Task.run(Task.scala:131)
- at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
- at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
- at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
- at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
- at java.base/java.lang.Thread.run(Thread.java:829)
- 24/05/19 13:23:02 ERROR TaskSetManager: Task 0 in stage 1.0 failed 1 times; aborting job
- 24/05/19 13:23:02 INFO TaskSchedulerImpl: Removed TaskSet 1.0, whose tasks have all completed, from pool
- 24/05/19 13:23:02 INFO TaskSchedulerImpl: Cancelling stage 1
- 24/05/19 13:23:02 INFO TaskSchedulerImpl: Killing all running tasks in stage 1: Stage cancelled
- 24/05/19 13:23:02 INFO DAGScheduler: ResultStage 1 (runJob at EsSparkSQL.scala:103) failed in 84.345 s due to Job aborted due to stage failure: Task 0 in stage 1.0 failed 1 times, most recent failure: Lost task 0.0 in stage 1.0 (TID 1) (Lenovo executor driver): org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException: Connection error (check network and/or proxy settings)- all nodes failed; tried [[172.19.0.2:9200]]
- at org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:160)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:442)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:438)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:398)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:402)
- at org.elasticsearch.hadoop.rest.RestClient.get(RestClient.java:178)
- at org.elasticsearch.hadoop.rest.request.GetAliasesRequestBuilder.execute(GetAliasesRequestBuilder.java:68)
- at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:620)
- at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:71)
- at org.elasticsearch.spark.sql.EsSparkSQL$.$anonfun$saveToEs$1(EsSparkSQL.scala:103)
- at org.elasticsearch.spark.sql.EsSparkSQL$.$anonfun$saveToEs$1$adapted(EsSparkSQL.scala:103)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
- at org.apache.spark.scheduler.Task.run(Task.scala:131)
- at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
- at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
- at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
- at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
- at java.base/java.lang.Thread.run(Thread.java:829)
- Driver stacktrace:
- 24/05/19 13:23:02 INFO DAGScheduler: Job 1 failed: runJob at EsSparkSQL.scala:103, took 84.348645 s
- org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 1.0 failed 1 times, most recent failure: Lost task 0.0 in stage 1.0 (TID 1) (Lenovo executor driver): org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException: Connection error (check network and/or proxy settings)- all nodes failed; tried [[172.19.0.2:9200]]
- at org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:160)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:442)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:438)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:398)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:402)
- at org.elasticsearch.hadoop.rest.RestClient.get(RestClient.java:178)
- at org.elasticsearch.hadoop.rest.request.GetAliasesRequestBuilder.execute(GetAliasesRequestBuilder.java:68)
- at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:620)
- at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:71)
- at org.elasticsearch.spark.sql.EsSparkSQL$.$anonfun$saveToEs$1(EsSparkSQL.scala:103)
- at org.elasticsearch.spark.sql.EsSparkSQL$.$anonfun$saveToEs$1$adapted(EsSparkSQL.scala:103)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
- at org.apache.spark.scheduler.Task.run(Task.scala:131)
- at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
- at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
- at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
- at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
- at java.base/java.lang.Thread.run(Thread.java:829)
- Driver stacktrace:
- at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2258)
- at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2207)
- at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2206)
- at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
- at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
- at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
- at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2206)
- at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1079)
- at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1079)
- at scala.Option.foreach(Option.scala:407)
- at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1079)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2445)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2387)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2376)
- at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
- at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:868)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2196)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2217)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2249)
- at org.elasticsearch.spark.sql.EsSparkSQL$.saveToEs(EsSparkSQL.scala:103)
- at org.elasticsearch.spark.sql.ElasticsearchRelation.insert(DefaultSource.scala:629)
- at org.elasticsearch.spark.sql.DefaultSource.createRelation(DefaultSource.scala:107)
- at org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:46)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
- at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:90)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:180)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:218)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:215)
- at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:176)
- at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:132)
- at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:131)
- at org.apache.spark.sql.DataFrameWriter.$anonfun$runCommand$1(DataFrameWriter.scala:989)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
- at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
- at org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:989)
- at org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:438)
- at org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:415)
- at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:301)
- at biz.svyatoslav.learning.bigdata.elasticsearch.SparkJavaElasticStreamTest.main(SparkJavaElasticStreamTest.java:38)
- Caused by: org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException: Connection error (check network and/or proxy settings)- all nodes failed; tried [[172.19.0.2:9200]]
- at org.elasticsearch.hadoop.rest.NetworkClient.execute(NetworkClient.java:160)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:442)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:438)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:398)
- at org.elasticsearch.hadoop.rest.RestClient.execute(RestClient.java:402)
- at org.elasticsearch.hadoop.rest.RestClient.get(RestClient.java:178)
- at org.elasticsearch.hadoop.rest.request.GetAliasesRequestBuilder.execute(GetAliasesRequestBuilder.java:68)
- at org.elasticsearch.hadoop.rest.RestService.createWriter(RestService.java:620)
- at org.elasticsearch.spark.rdd.EsRDDWriter.write(EsRDDWriter.scala:71)
- at org.elasticsearch.spark.sql.EsSparkSQL$.$anonfun$saveToEs$1(EsSparkSQL.scala:103)
- at org.elasticsearch.spark.sql.EsSparkSQL$.$anonfun$saveToEs$1$adapted(EsSparkSQL.scala:103)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
- at org.apache.spark.scheduler.Task.run(Task.scala:131)
- at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
- at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
- at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
- at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
- at java.base/java.lang.Thread.run(Thread.java:829)
- 24/05/19 13:23:02 INFO SparkContext: Invoking stop() from shutdown hook
- 24/05/19 13:23:02 INFO SparkUI: Stopped Spark web UI at http://Lenovo:4040
- 24/05/19 13:23:02 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
- 24/05/19 13:23:02 INFO MemoryStore: MemoryStore cleared
- 24/05/19 13:23:02 INFO BlockManager: BlockManager stopped
- 24/05/19 13:23:02 INFO BlockManagerMaster: BlockManagerMaster stopped
- 24/05/19 13:23:02 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
- 24/05/19 13:23:02 INFO SparkContext: Successfully stopped SparkContext
- 24/05/19 13:23:02 INFO ShutdownHookManager: Shutdown hook called
- 24/05/19 13:23:02 INFO ShutdownHookManager: Deleting directory C:\Users\Lenovo\AppData\Local\Temp\spark-bea50fa0-d01b-46c9-865c-b815d505e006
- Disconnected from the target VM, address: '127.0.0.1:55141', transport: 'socket'
- Process finished with exit code 0
Add Comment
Please, Sign In to add comment