Skip to content

Instantly share code, notes, and snippets.

@Kontinuation
Created September 18, 2021 03:56
Show Gist options
  • Save Kontinuation/056ae64d3ecfd5a03a1dd2b45a6dd12f to your computer and use it in GitHub Desktop.
Save Kontinuation/056ae64d3ecfd5a03a1dd2b45a6dd12f to your computer and use it in GitHub Desktop.
Spatial join fails after merging commit a0ed49 ([SEDONA-28] shape&wkb ser de)
21/09/18 11:51:31 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
21/09/18 11:51:31 INFO SparkContext: Running Spark version 3.1.2
21/09/18 11:51:31 INFO ResourceUtils: ==============================================================
21/09/18 11:51:31 INFO ResourceUtils: No custom resources configured for spark.driver.
21/09/18 11:51:31 INFO ResourceUtils: ==============================================================
21/09/18 11:51:31 INFO SparkContext: Submitted application: SedonaSQL-demo
21/09/18 11:51:31 INFO ResourceProfile: Default ResourceProfile created, executor resources: Map(cores -> name: cores, amount: 1, script: , vendor: , memory -> name: memory, amount: 1024, script: , vendor: , offHeap -> name: offHeap, amount: 0, script: , vendor: ), task resources: Map(cpus -> name: cpus, amount: 1.0)
21/09/18 11:51:31 INFO ResourceProfile: Limiting resource is cpu
21/09/18 11:51:31 INFO ResourceProfileManager: Added ResourceProfile id: 0
21/09/18 11:51:31 INFO SecurityManager: Changing view acls to: kontinuation
21/09/18 11:51:31 INFO SecurityManager: Changing modify acls to: kontinuation
21/09/18 11:51:31 INFO SecurityManager: Changing view acls groups to:
21/09/18 11:51:31 INFO SecurityManager: Changing modify acls groups to:
21/09/18 11:51:31 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(kontinuation); groups with view permissions: Set(); users with modify permissions: Set(kontinuation); groups with modify permissions: Set()
21/09/18 11:51:31 INFO Utils: Successfully started service 'sparkDriver' on port 35501.
21/09/18 11:51:31 INFO SparkEnv: Registering MapOutputTracker
21/09/18 11:51:31 INFO SparkEnv: Registering BlockManagerMaster
21/09/18 11:51:31 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
21/09/18 11:51:31 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
21/09/18 11:51:31 INFO SparkEnv: Registering BlockManagerMasterHeartbeat
21/09/18 11:51:31 INFO DiskBlockManager: Created local directory at /tmp/blockmgr-46091ea7-98e3-405b-b461-93b295962c72
21/09/18 11:51:31 INFO MemoryStore: MemoryStore started with capacity 366.3 MiB
21/09/18 11:51:32 INFO SparkEnv: Registering OutputCommitCoordinator
21/09/18 11:51:32 INFO Utils: Successfully started service 'SparkUI' on port 4040.
21/09/18 11:51:32 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at http://kontinuation:4040
21/09/18 11:51:32 INFO SparkContext: Added JAR file:/home/kontinuation/documents/geospatial/experi/sedona-test/target/sedona-test_2.12-1.0.0-SNAPSHOT-shaded.jar at spark://kontinuation:35501/jars/sedona-test_2.12-1.0.0-SNAPSHOT-shaded.jar with timestamp 1631937091411
21/09/18 11:51:32 INFO StandaloneAppClient$ClientEndpoint: Connecting to master spark://kontinuation:7077...
21/09/18 11:51:32 INFO TransportClientFactory: Successfully created connection to kontinuation/172.25.87.59:7077 after 29 ms (0 ms spent in bootstraps)
21/09/18 11:51:32 INFO StandaloneSchedulerBackend: Connected to Spark cluster with app ID app-20210918115132-0003
21/09/18 11:51:32 INFO StandaloneAppClient$ClientEndpoint: Executor added: app-20210918115132-0003/0 on worker-20210918105325-172.25.87.59-40855 (172.25.87.59:40855) with 4 core(s)
21/09/18 11:51:32 INFO StandaloneSchedulerBackend: Granted executor ID app-20210918115132-0003/0 on hostPort 172.25.87.59:40855 with 4 core(s), 1024.0 MiB RAM
21/09/18 11:51:32 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 44191.
21/09/18 11:51:32 INFO NettyBlockTransferService: Server created on kontinuation:44191
21/09/18 11:51:32 INFO BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
21/09/18 11:51:32 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(driver, kontinuation, 44191, None)
21/09/18 11:51:32 INFO BlockManagerMasterEndpoint: Registering block manager kontinuation:44191 with 366.3 MiB RAM, BlockManagerId(driver, kontinuation, 44191, None)
21/09/18 11:51:32 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(driver, kontinuation, 44191, None)
21/09/18 11:51:32 INFO BlockManager: Initialized BlockManager: BlockManagerId(driver, kontinuation, 44191, None)
21/09/18 11:51:32 INFO StandaloneAppClient$ClientEndpoint: Executor updated: app-20210918115132-0003/0 is now RUNNING
21/09/18 11:51:33 INFO StandaloneSchedulerBackend: SchedulerBackend is ready for scheduling beginning after reached minRegisteredResourcesRatio: 0.0
21/09/18 11:51:33 INFO SharedState: Setting hive.metastore.warehouse.dir ('null') to the value of spark.sql.warehouse.dir ('file:/home/kontinuation/documents/geospatial/experi/sedona-test/spark-warehouse').
21/09/18 11:51:33 INFO SharedState: Warehouse path is 'file:/home/kontinuation/documents/geospatial/experi/sedona-test/spark-warehouse'.
21/09/18 11:51:35 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Registered executor NettyRpcEndpointRef(spark-client://Executor) (172.25.87.59:53362) with ID 0, ResourceProfileId 0
21/09/18 11:51:35 INFO BlockManagerMasterEndpoint: Registering block manager 172.25.87.59:44483 with 366.3 MiB RAM, BlockManagerId(0, 172.25.87.59, 44483, None)
Sedona Configuration:
useIndex: false
indexType: QUADTREE
serializerType: SHAPE
joinSparitionDominantSide: LEFT
joinBuildSide: LEFT
joinApproximateTotalCount: -1
datasetBoundary: Env[0.0 : 0.0, 0.0 : 0.0]
fallbackPartitionNum: -1
joinGridType: KDBTREE
21/09/18 11:51:36 INFO InMemoryFileIndex: It took 34 ms to list leaf files for 1 paths.
21/09/18 11:51:36 INFO InMemoryFileIndex: It took 2 ms to list leaf files for 1 paths.
21/09/18 11:51:37 INFO FileSourceStrategy: Pushed Filters:
21/09/18 11:51:37 INFO FileSourceStrategy: Post-Scan Filters: (length(trim(value#6, None)) > 0)
21/09/18 11:51:37 INFO FileSourceStrategy: Output Data Schema: struct<value: string>
21/09/18 11:51:38 INFO CodeGenerator: Code generated in 255.211895 ms
21/09/18 11:51:38 INFO MemoryStore: Block broadcast_0 stored as values in memory (estimated size 286.9 KiB, free 366.0 MiB)
21/09/18 11:51:38 INFO SedonaWKBKryoRegistrator: Registering custom serializers for geometry types
21/09/18 11:51:38 INFO SedonaVizKryoRegistrator: Registering custom serializers for visualization related types
21/09/18 11:51:38 INFO MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 24.2 KiB, free 366.0 MiB)
21/09/18 11:51:38 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory on kontinuation:44191 (size: 24.2 KiB, free: 366.3 MiB)
21/09/18 11:51:38 INFO SparkContext: Created broadcast 0 from load at ExperiSedona.scala:53
21/09/18 11:51:38 INFO FileSourceScanExec: Planning scan with bin packing, max size: 4194304 bytes, open cost is considered as scanning 4194304 bytes.
21/09/18 11:51:38 INFO SparkContext: Starting job: load at ExperiSedona.scala:53
21/09/18 11:51:38 INFO DAGScheduler: Got job 0 (load at ExperiSedona.scala:53) with 1 output partitions
21/09/18 11:51:38 INFO DAGScheduler: Final stage: ResultStage 0 (load at ExperiSedona.scala:53)
21/09/18 11:51:38 INFO DAGScheduler: Parents of final stage: List()
21/09/18 11:51:38 INFO DAGScheduler: Missing parents: List()
21/09/18 11:51:38 INFO DAGScheduler: Submitting ResultStage 0 (MapPartitionsRDD[3] at load at ExperiSedona.scala:53), which has no missing parents
21/09/18 11:51:39 INFO MemoryStore: Block broadcast_1 stored as values in memory (estimated size 10.8 KiB, free 366.0 MiB)
21/09/18 11:51:39 INFO MemoryStore: Block broadcast_1_piece0 stored as bytes in memory (estimated size 5.4 KiB, free 366.0 MiB)
21/09/18 11:51:39 INFO BlockManagerInfo: Added broadcast_1_piece0 in memory on kontinuation:44191 (size: 5.4 KiB, free: 366.3 MiB)
21/09/18 11:51:39 INFO SparkContext: Created broadcast 1 from broadcast at DAGScheduler.scala:1388
21/09/18 11:51:39 INFO DAGScheduler: Submitting 1 missing tasks from ResultStage 0 (MapPartitionsRDD[3] at load at ExperiSedona.scala:53) (first 15 tasks are for partitions Vector(0))
21/09/18 11:51:39 INFO TaskSchedulerImpl: Adding task set 0.0 with 1 tasks resource profile 0
21/09/18 11:51:39 INFO TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0) (172.25.87.59, executor 0, partition 0, PROCESS_LOCAL, 4925 bytes) taskResourceAssignments Map()
21/09/18 11:51:39 INFO BlockManagerInfo: Added broadcast_1_piece0 in memory on 172.25.87.59:44483 (size: 5.4 KiB, free: 366.3 MiB)
21/09/18 11:51:40 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory on 172.25.87.59:44483 (size: 24.2 KiB, free: 366.3 MiB)
21/09/18 11:51:41 INFO TaskSetManager: Finished task 0.0 in stage 0.0 (TID 0) in 2236 ms on 172.25.87.59 (executor 0) (1/1)
21/09/18 11:51:41 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose tasks have all completed, from pool
21/09/18 11:51:41 INFO DAGScheduler: ResultStage 0 (load at ExperiSedona.scala:53) finished in 2.364 s
21/09/18 11:51:41 INFO DAGScheduler: Job 0 is finished. Cancelling potential speculative or zombie tasks for this job
21/09/18 11:51:41 INFO TaskSchedulerImpl: Killing all running tasks in stage 0: Stage finished
21/09/18 11:51:41 INFO DAGScheduler: Job 0 finished: load at ExperiSedona.scala:53, took 2.417185 s
21/09/18 11:51:41 INFO CodeGenerator: Code generated in 13.549909 ms
21/09/18 11:51:41 INFO FileSourceStrategy: Pushed Filters:
21/09/18 11:51:41 INFO FileSourceStrategy: Post-Scan Filters:
21/09/18 11:51:41 INFO FileSourceStrategy: Output Data Schema: struct<value: string>
21/09/18 11:51:41 INFO MemoryStore: Block broadcast_2 stored as values in memory (estimated size 286.9 KiB, free 365.7 MiB)
21/09/18 11:51:41 INFO MemoryStore: Block broadcast_2_piece0 stored as bytes in memory (estimated size 24.2 KiB, free 365.7 MiB)
21/09/18 11:51:41 INFO BlockManagerInfo: Added broadcast_2_piece0 in memory on kontinuation:44191 (size: 24.2 KiB, free: 366.2 MiB)
21/09/18 11:51:41 INFO SparkContext: Created broadcast 2 from load at ExperiSedona.scala:53
21/09/18 11:51:41 INFO FileSourceScanExec: Planning scan with bin packing, max size: 4194304 bytes, open cost is considered as scanning 4194304 bytes.
21/09/18 11:51:41 INFO InMemoryFileIndex: It took 1 ms to list leaf files for 1 paths.
21/09/18 11:51:41 INFO InMemoryFileIndex: It took 1 ms to list leaf files for 1 paths.
21/09/18 11:51:41 INFO FileSourceStrategy: Pushed Filters:
21/09/18 11:51:41 INFO FileSourceStrategy: Post-Scan Filters: (length(trim(value#30, None)) > 0)
21/09/18 11:51:41 INFO FileSourceStrategy: Output Data Schema: struct<value: string>
21/09/18 11:51:41 INFO MemoryStore: Block broadcast_3 stored as values in memory (estimated size 286.9 KiB, free 365.4 MiB)
21/09/18 11:51:41 INFO MemoryStore: Block broadcast_3_piece0 stored as bytes in memory (estimated size 24.2 KiB, free 365.4 MiB)
21/09/18 11:51:41 INFO BlockManagerInfo: Added broadcast_3_piece0 in memory on kontinuation:44191 (size: 24.2 KiB, free: 366.2 MiB)
21/09/18 11:51:41 INFO SparkContext: Created broadcast 3 from load at ExperiSedona.scala:61
21/09/18 11:51:41 INFO FileSourceScanExec: Planning scan with bin packing, max size: 4194304 bytes, open cost is considered as scanning 4194304 bytes.
21/09/18 11:51:41 INFO SparkContext: Starting job: load at ExperiSedona.scala:61
21/09/18 11:51:41 INFO DAGScheduler: Got job 1 (load at ExperiSedona.scala:61) with 1 output partitions
21/09/18 11:51:41 INFO DAGScheduler: Final stage: ResultStage 1 (load at ExperiSedona.scala:61)
21/09/18 11:51:41 INFO DAGScheduler: Parents of final stage: List()
21/09/18 11:51:41 INFO DAGScheduler: Missing parents: List()
21/09/18 11:51:41 INFO DAGScheduler: Submitting ResultStage 1 (MapPartitionsRDD[13] at load at ExperiSedona.scala:61), which has no missing parents
21/09/18 11:51:41 INFO MemoryStore: Block broadcast_4 stored as values in memory (estimated size 10.8 KiB, free 365.4 MiB)
21/09/18 11:51:41 INFO MemoryStore: Block broadcast_4_piece0 stored as bytes in memory (estimated size 5.4 KiB, free 365.4 MiB)
21/09/18 11:51:41 INFO BlockManagerInfo: Added broadcast_4_piece0 in memory on kontinuation:44191 (size: 5.4 KiB, free: 366.2 MiB)
21/09/18 11:51:41 INFO SparkContext: Created broadcast 4 from broadcast at DAGScheduler.scala:1388
21/09/18 11:51:41 INFO DAGScheduler: Submitting 1 missing tasks from ResultStage 1 (MapPartitionsRDD[13] at load at ExperiSedona.scala:61) (first 15 tasks are for partitions Vector(0))
21/09/18 11:51:41 INFO TaskSchedulerImpl: Adding task set 1.0 with 1 tasks resource profile 0
21/09/18 11:51:42 INFO TaskSetManager: Starting task 0.0 in stage 1.0 (TID 1) (172.25.87.59, executor 0, partition 0, PROCESS_LOCAL, 4921 bytes) taskResourceAssignments Map()
21/09/18 11:51:42 INFO BlockManagerInfo: Added broadcast_4_piece0 in memory on 172.25.87.59:44483 (size: 5.4 KiB, free: 366.3 MiB)
21/09/18 11:51:42 INFO BlockManagerInfo: Added broadcast_3_piece0 in memory on 172.25.87.59:44483 (size: 24.2 KiB, free: 366.2 MiB)
21/09/18 11:51:42 INFO TaskSetManager: Finished task 0.0 in stage 1.0 (TID 1) in 135 ms on 172.25.87.59 (executor 0) (1/1)
21/09/18 11:51:42 INFO TaskSchedulerImpl: Removed TaskSet 1.0, whose tasks have all completed, from pool
21/09/18 11:51:42 INFO DAGScheduler: ResultStage 1 (load at ExperiSedona.scala:61) finished in 0.144 s
21/09/18 11:51:42 INFO DAGScheduler: Job 1 is finished. Cancelling potential speculative or zombie tasks for this job
21/09/18 11:51:42 INFO TaskSchedulerImpl: Killing all running tasks in stage 1: Stage finished
21/09/18 11:51:42 INFO DAGScheduler: Job 1 finished: load at ExperiSedona.scala:61, took 0.148971 s
21/09/18 11:51:42 INFO FileSourceStrategy: Pushed Filters:
21/09/18 11:51:42 INFO FileSourceStrategy: Post-Scan Filters:
21/09/18 11:51:42 INFO FileSourceStrategy: Output Data Schema: struct<value: string>
21/09/18 11:51:42 INFO MemoryStore: Block broadcast_5 stored as values in memory (estimated size 286.9 KiB, free 365.1 MiB)
21/09/18 11:51:42 INFO MemoryStore: Block broadcast_5_piece0 stored as bytes in memory (estimated size 24.2 KiB, free 365.1 MiB)
21/09/18 11:51:42 INFO BlockManagerInfo: Added broadcast_5_piece0 in memory on kontinuation:44191 (size: 24.2 KiB, free: 366.2 MiB)
21/09/18 11:51:42 INFO SparkContext: Created broadcast 5 from load at ExperiSedona.scala:61
21/09/18 11:51:42 INFO FileSourceScanExec: Planning scan with bin packing, max size: 4194304 bytes, open cost is considered as scanning 4194304 bytes.
21/09/18 11:51:42 INFO JoinQueryDetector: Planning spatial join for ST_Contains relationship
21/09/18 11:51:42 INFO FileSourceStrategy: Pushed Filters:
21/09/18 11:51:42 INFO FileSourceStrategy: Post-Scan Filters:
21/09/18 11:51:42 INFO FileSourceStrategy: Output Data Schema: struct<_c0: string, _c1: string>
21/09/18 11:51:42 INFO FileSourceStrategy: Pushed Filters:
21/09/18 11:51:42 INFO FileSourceStrategy: Post-Scan Filters:
21/09/18 11:51:42 INFO FileSourceStrategy: Output Data Schema: struct<_c1: string>
== Physical Plan ==
*(3) Sort [cnt#54L DESC NULLS LAST], true, 0
+- Exchange rangepartitioning(cnt#54L DESC NULLS LAST, 200), ENSURE_REQUIREMENTS, [id=#93]
+- *(2) HashAggregate(keys=[name#26], functions=[count(1)])
+- Exchange hashpartitioning(name#26, 200), ENSURE_REQUIREMENTS, [id=#89]
+- *(1) HashAggregate(keys=[name#26], functions=[partial_count(1)])
+- *(1) Project [name#26]
+- RangeJoin geom#27: geometry, geom#51: geometry, false
:- Project [_c0#22 AS name#26, st_geomfromwkt(_c1#23) AS geom#27]
: +- FileScan csv [_c0#22,_c1#23] Batched: false, DataFilters: [], Format: CSV, Location: InMemoryFileIndex[file:/home/kontinuation/documents/geospatial/experi/sedona-test/data/china_prov..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<_c0:string,_c1:string>
+- Project [st_geomfromwkt(_c1#47) AS geom#51]
+- FileScan csv [_c1#47] Batched: false, DataFilters: [], Format: CSV, Location: InMemoryFileIndex[file:/home/kontinuation/documents/geospatial/experi/sedona-test/data/ubm_trip_r..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<_c1:string>
21/09/18 11:51:42 INFO JoinQueryDetector: Planning spatial join for ST_Contains relationship
21/09/18 11:51:42 INFO FileSourceStrategy: Pushed Filters:
21/09/18 11:51:42 INFO FileSourceStrategy: Post-Scan Filters:
21/09/18 11:51:42 INFO FileSourceStrategy: Output Data Schema: struct<_c0: string, _c1: string>
21/09/18 11:51:42 INFO FileSourceStrategy: Pushed Filters:
21/09/18 11:51:42 INFO FileSourceStrategy: Post-Scan Filters:
21/09/18 11:51:42 INFO FileSourceStrategy: Output Data Schema: struct<_c1: string>
21/09/18 11:51:42 INFO CodeGenerator: Code generated in 14.883032 ms
21/09/18 11:51:42 INFO CodeGenerator: Code generated in 30.195449 ms
21/09/18 11:51:42 INFO CodeGenerator: Code generated in 44.536571 ms
21/09/18 11:51:42 INFO MemoryStore: Block broadcast_6 stored as values in memory (estimated size 286.8 KiB, free 364.8 MiB)
21/09/18 11:51:42 INFO MemoryStore: Block broadcast_6_piece0 stored as bytes in memory (estimated size 24.3 KiB, free 364.7 MiB)
21/09/18 11:51:42 INFO BlockManagerInfo: Added broadcast_6_piece0 in memory on kontinuation:44191 (size: 24.3 KiB, free: 366.2 MiB)
21/09/18 11:51:42 INFO SparkContext: Created broadcast 6 from show at ExperiSedona.scala:73
21/09/18 11:51:42 INFO FileSourceScanExec: Planning scan with bin packing, max size: 4194304 bytes, open cost is considered as scanning 4194304 bytes.
21/09/18 11:51:42 INFO MemoryStore: Block broadcast_7 stored as values in memory (estimated size 286.8 KiB, free 364.5 MiB)
21/09/18 11:51:42 INFO MemoryStore: Block broadcast_7_piece0 stored as bytes in memory (estimated size 24.3 KiB, free 364.4 MiB)
21/09/18 11:51:42 INFO BlockManagerInfo: Added broadcast_7_piece0 in memory on kontinuation:44191 (size: 24.3 KiB, free: 366.1 MiB)
21/09/18 11:51:42 INFO SparkContext: Created broadcast 7 from show at ExperiSedona.scala:73
21/09/18 11:51:42 INFO FileSourceScanExec: Planning scan with bin packing, max size: 4194304 bytes, open cost is considered as scanning 4194304 bytes.
21/09/18 11:51:42 INFO SparkContext: Starting job: aggregate at SpatialRDD.java:481
21/09/18 11:51:42 INFO DAGScheduler: Got job 2 (aggregate at SpatialRDD.java:481) with 2 output partitions
21/09/18 11:51:42 INFO DAGScheduler: Final stage: ResultStage 2 (aggregate at SpatialRDD.java:481)
21/09/18 11:51:42 INFO DAGScheduler: Parents of final stage: List()
21/09/18 11:51:42 INFO DAGScheduler: Missing parents: List()
21/09/18 11:51:42 INFO DAGScheduler: Submitting ResultStage 2 (MapPartitionsRDD[26] at show at ExperiSedona.scala:73), which has no missing parents
21/09/18 11:51:42 INFO MemoryStore: Block broadcast_8 stored as values in memory (estimated size 14.5 KiB, free 364.4 MiB)
21/09/18 11:51:42 INFO MemoryStore: Block broadcast_8_piece0 stored as bytes in memory (estimated size 7.6 KiB, free 364.4 MiB)
21/09/18 11:51:42 INFO BlockManagerInfo: Added broadcast_8_piece0 in memory on kontinuation:44191 (size: 7.6 KiB, free: 366.1 MiB)
21/09/18 11:51:42 INFO SparkContext: Created broadcast 8 from broadcast at DAGScheduler.scala:1388
21/09/18 11:51:42 INFO DAGScheduler: Submitting 2 missing tasks from ResultStage 2 (MapPartitionsRDD[26] at show at ExperiSedona.scala:73) (first 15 tasks are for partitions Vector(0, 1))
21/09/18 11:51:42 INFO TaskSchedulerImpl: Adding task set 2.0 with 2 tasks resource profile 0
21/09/18 11:51:42 INFO TaskSetManager: Starting task 0.0 in stage 2.0 (TID 2) (172.25.87.59, executor 0, partition 0, PROCESS_LOCAL, 4925 bytes) taskResourceAssignments Map()
21/09/18 11:51:42 INFO TaskSetManager: Starting task 1.0 in stage 2.0 (TID 3) (172.25.87.59, executor 0, partition 1, PROCESS_LOCAL, 4925 bytes) taskResourceAssignments Map()
21/09/18 11:51:42 INFO BlockManagerInfo: Added broadcast_8_piece0 in memory on 172.25.87.59:44483 (size: 7.6 KiB, free: 366.2 MiB)
21/09/18 11:51:43 INFO BlockManagerInfo: Added broadcast_6_piece0 in memory on 172.25.87.59:44483 (size: 24.3 KiB, free: 366.2 MiB)
21/09/18 11:51:43 WARN TaskSetManager: Lost task 0.0 in stage 2.0 (TID 2) (172.25.87.59 executor 0): java.lang.NoClassDefFoundError: Could not initialize class org.apache.spark.sql.sedona_sql.package$
at org.apache.spark.sql.sedona_sql.expressions.ST_GeomFromWKT.eval(Constructors.scala:131)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection.apply(Unknown Source)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection.apply(Unknown Source)
at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)
at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)
at scala.collection.Iterator.foreach(Iterator.scala:941)
at scala.collection.Iterator.foreach$(Iterator.scala:941)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1429)
at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:162)
at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:160)
at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1429)
at scala.collection.TraversableOnce.aggregate(TraversableOnce.scala:219)
at scala.collection.TraversableOnce.aggregate$(TraversableOnce.scala:219)
at scala.collection.AbstractIterator.aggregate(Iterator.scala:1429)
at org.apache.spark.rdd.RDD.$anonfun$aggregate$2(RDD.scala:1207)
at org.apache.spark.SparkContext.$anonfun$runJob$6(SparkContext.scala:2290)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
21/09/18 11:51:43 WARN TaskSetManager: Lost task 1.0 in stage 2.0 (TID 3) (172.25.87.59 executor 0): java.lang.ExceptionInInitializerError
at org.apache.spark.sql.sedona_sql.expressions.ST_GeomFromWKT.eval(Constructors.scala:131)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection.apply(Unknown Source)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection.apply(Unknown Source)
at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)
at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)
at scala.collection.Iterator.foreach(Iterator.scala:941)
at scala.collection.Iterator.foreach$(Iterator.scala:941)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1429)
at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:162)
at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:160)
at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1429)
at scala.collection.TraversableOnce.aggregate(TraversableOnce.scala:219)
at scala.collection.TraversableOnce.aggregate$(TraversableOnce.scala:219)
at scala.collection.AbstractIterator.aggregate(Iterator.scala:1429)
at org.apache.spark.rdd.RDD.$anonfun$aggregate$2(RDD.scala:1207)
at org.apache.spark.SparkContext.$anonfun$runJob$6(SparkContext.scala:2290)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.spark.SparkException: There is no active SparkContext. Hence, cannot create SedonaSerializer
at org.apache.spark.sql.sedona_sql.package$.<init>(package.scala:31)
at org.apache.spark.sql.sedona_sql.package$.<clinit>(package.scala)
... 24 more
21/09/18 11:51:43 INFO TaskSetManager: Starting task 1.1 in stage 2.0 (TID 4) (172.25.87.59, executor 0, partition 1, PROCESS_LOCAL, 4925 bytes) taskResourceAssignments Map()
21/09/18 11:51:43 INFO TaskSetManager: Starting task 0.1 in stage 2.0 (TID 5) (172.25.87.59, executor 0, partition 0, PROCESS_LOCAL, 4925 bytes) taskResourceAssignments Map()
21/09/18 11:51:43 INFO TaskSetManager: Lost task 0.1 in stage 2.0 (TID 5) on 172.25.87.59, executor 0: java.lang.NoClassDefFoundError (Could not initialize class org.apache.spark.sql.sedona_sql.package$) [duplicate 1]
21/09/18 11:51:43 INFO TaskSetManager: Starting task 0.2 in stage 2.0 (TID 6) (172.25.87.59, executor 0, partition 0, PROCESS_LOCAL, 4925 bytes) taskResourceAssignments Map()
21/09/18 11:51:43 INFO TaskSetManager: Lost task 1.1 in stage 2.0 (TID 4) on 172.25.87.59, executor 0: java.lang.NoClassDefFoundError (Could not initialize class org.apache.spark.sql.sedona_sql.package$) [duplicate 2]
21/09/18 11:51:43 INFO TaskSetManager: Starting task 1.2 in stage 2.0 (TID 7) (172.25.87.59, executor 0, partition 1, PROCESS_LOCAL, 4925 bytes) taskResourceAssignments Map()
21/09/18 11:51:43 INFO TaskSetManager: Lost task 0.2 in stage 2.0 (TID 6) on 172.25.87.59, executor 0: java.lang.NoClassDefFoundError (Could not initialize class org.apache.spark.sql.sedona_sql.package$) [duplicate 3]
21/09/18 11:51:43 INFO TaskSetManager: Starting task 0.3 in stage 2.0 (TID 8) (172.25.87.59, executor 0, partition 0, PROCESS_LOCAL, 4925 bytes) taskResourceAssignments Map()
21/09/18 11:51:43 INFO TaskSetManager: Lost task 1.2 in stage 2.0 (TID 7) on 172.25.87.59, executor 0: java.lang.NoClassDefFoundError (Could not initialize class org.apache.spark.sql.sedona_sql.package$) [duplicate 4]
21/09/18 11:51:43 INFO TaskSetManager: Starting task 1.3 in stage 2.0 (TID 9) (172.25.87.59, executor 0, partition 1, PROCESS_LOCAL, 4925 bytes) taskResourceAssignments Map()
21/09/18 11:51:43 INFO TaskSetManager: Lost task 0.3 in stage 2.0 (TID 8) on 172.25.87.59, executor 0: java.lang.NoClassDefFoundError (Could not initialize class org.apache.spark.sql.sedona_sql.package$) [duplicate 5]
21/09/18 11:51:43 ERROR TaskSetManager: Task 0 in stage 2.0 failed 4 times; aborting job
21/09/18 11:51:43 INFO TaskSchedulerImpl: Cancelling stage 2
21/09/18 11:51:43 INFO TaskSchedulerImpl: Killing all running tasks in stage 2: Stage cancelled
21/09/18 11:51:43 INFO TaskSchedulerImpl: Stage 2 was cancelled
21/09/18 11:51:43 INFO DAGScheduler: ResultStage 2 (aggregate at SpatialRDD.java:481) failed in 0.967 s due to Job aborted due to stage failure: Task 0 in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in stage 2.0 (TID 8) (172.25.87.59 executor 0): java.lang.NoClassDefFoundError: Could not initialize class org.apache.spark.sql.sedona_sql.package$
at org.apache.spark.sql.sedona_sql.expressions.ST_GeomFromWKT.eval(Constructors.scala:131)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection.apply(Unknown Source)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection.apply(Unknown Source)
at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)
at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)
at scala.collection.Iterator.foreach(Iterator.scala:941)
at scala.collection.Iterator.foreach$(Iterator.scala:941)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1429)
at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:162)
at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:160)
at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1429)
at scala.collection.TraversableOnce.aggregate(TraversableOnce.scala:219)
at scala.collection.TraversableOnce.aggregate$(TraversableOnce.scala:219)
at scala.collection.AbstractIterator.aggregate(Iterator.scala:1429)
at org.apache.spark.rdd.RDD.$anonfun$aggregate$2(RDD.scala:1207)
at org.apache.spark.SparkContext.$anonfun$runJob$6(SparkContext.scala:2290)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Driver stacktrace:
21/09/18 11:51:43 INFO DAGScheduler: Job 2 failed: aggregate at SpatialRDD.java:481, took 0.973420 s
21/09/18 11:51:43 INFO TaskSetManager: Lost task 1.3 in stage 2.0 (TID 9) on 172.25.87.59, executor 0: java.lang.NoClassDefFoundError (Could not initialize class org.apache.spark.sql.sedona_sql.package$) [duplicate 6]
21/09/18 11:51:43 INFO TaskSchedulerImpl: Removed TaskSet 2.0, whose tasks have all completed, from pool
21/09/18 11:51:43 INFO SparkUI: Stopped Spark web UI at http://kontinuation:4040
21/09/18 11:51:43 INFO StandaloneSchedulerBackend: Shutting down all executors
21/09/18 11:51:43 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Asking each executor to shut down
21/09/18 11:51:43 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
21/09/18 11:51:43 INFO MemoryStore: MemoryStore cleared
21/09/18 11:51:43 INFO BlockManager: BlockManager stopped
21/09/18 11:51:43 INFO BlockManagerMaster: BlockManagerMaster stopped
21/09/18 11:51:43 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
21/09/18 11:51:43 INFO SparkContext: Successfully stopped SparkContext
Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in stage 2.0 (TID 8) (172.25.87.59 executor 0): java.lang.NoClassDefFoundError: Could not initialize class org.apache.spark.sql.sedona_sql.package$
at org.apache.spark.sql.sedona_sql.expressions.ST_GeomFromWKT.eval(Constructors.scala:131)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection.apply(Unknown Source)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection.apply(Unknown Source)
at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)
at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)
at scala.collection.Iterator.foreach(Iterator.scala:941)
at scala.collection.Iterator.foreach$(Iterator.scala:941)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1429)
at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:162)
at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:160)
at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1429)
at scala.collection.TraversableOnce.aggregate(TraversableOnce.scala:219)
at scala.collection.TraversableOnce.aggregate$(TraversableOnce.scala:219)
at scala.collection.AbstractIterator.aggregate(Iterator.scala:1429)
at org.apache.spark.rdd.RDD.$anonfun$aggregate$2(RDD.scala:1207)
at org.apache.spark.SparkContext.$anonfun$runJob$6(SparkContext.scala:2290)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2258)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2207)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2206)
at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2206)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1079)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1079)
at scala.Option.foreach(Option.scala:407)
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1079)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2445)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2387)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2376)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:868)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2196)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2291)
at org.apache.spark.rdd.RDD.$anonfun$aggregate$1(RDD.scala:1209)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:414)
at org.apache.spark.rdd.RDD.aggregate(RDD.scala:1202)
at org.apache.spark.api.java.JavaRDDLike.aggregate(JavaRDDLike.scala:427)
at org.apache.spark.api.java.JavaRDDLike.aggregate$(JavaRDDLike.scala:425)
at org.apache.spark.api.java.AbstractJavaRDDLike.aggregate(JavaRDDLike.scala:45)
at org.apache.sedona.core.spatialRDD.SpatialRDD.analyze(SpatialRDD.java:481)
at org.apache.spark.sql.sedona_sql.strategy.join.TraitJoinQueryExec.doExecute(TraitJoinQueryExec.scala:67)
at org.apache.spark.sql.sedona_sql.strategy.join.TraitJoinQueryExec.doExecute$(TraitJoinQueryExec.scala:52)
at org.apache.spark.sql.sedona_sql.strategy.join.RangeJoinExec.doExecute(RangeJoinExec.scala:37)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:180)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:218)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:215)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:176)
at org.apache.spark.sql.execution.InputAdapter.inputRDD(WholeStageCodegenExec.scala:525)
at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs(WholeStageCodegenExec.scala:453)
at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs$(WholeStageCodegenExec.scala:452)
at org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:496)
at org.apache.spark.sql.execution.ProjectExec.inputRDDs(basicPhysicalOperators.scala:50)
at org.apache.spark.sql.execution.aggregate.HashAggregateExec.inputRDDs(HashAggregateExec.scala:141)
at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:746)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:180)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:218)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:215)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:176)
at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.inputRDD$lzycompute(ShuffleExchangeExec.scala:118)
at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.inputRDD(ShuffleExchangeExec.scala:118)
at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.shuffleDependency$lzycompute(ShuffleExchangeExec.scala:151)
at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.shuffleDependency(ShuffleExchangeExec.scala:149)
at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.$anonfun$doExecute$1(ShuffleExchangeExec.scala:166)
at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:52)
at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.doExecute(ShuffleExchangeExec.scala:163)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:180)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:218)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:215)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:176)
at org.apache.spark.sql.execution.InputAdapter.inputRDD(WholeStageCodegenExec.scala:525)
at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs(WholeStageCodegenExec.scala:453)
at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs$(WholeStageCodegenExec.scala:452)
at org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:496)
at org.apache.spark.sql.execution.aggregate.HashAggregateExec.inputRDDs(HashAggregateExec.scala:141)
at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:746)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:180)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:218)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:215)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:176)
at org.apache.spark.sql.execution.TakeOrderedAndProjectExec.executeCollect(limit.scala:187)
at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:3696)
at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:2722)
at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3687)
at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)
at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3685)
at org.apache.spark.sql.Dataset.head(Dataset.scala:2722)
at org.apache.spark.sql.Dataset.take(Dataset.scala:2929)
at org.apache.spark.sql.Dataset.getRows(Dataset.scala:301)
at org.apache.spark.sql.Dataset.showString(Dataset.scala:338)
at org.apache.spark.sql.Dataset.show(Dataset.scala:825)
at org.apache.spark.sql.Dataset.show(Dataset.scala:784)
at org.apache.spark.sql.Dataset.show(Dataset.scala:793)
at ExperiSedona$.main(ExperiSedona.scala:73)
at ExperiSedona.main(ExperiSedona.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:951)
at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1039)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1048)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.NoClassDefFoundError: Could not initialize class org.apache.spark.sql.sedona_sql.package$
at org.apache.spark.sql.sedona_sql.expressions.ST_GeomFromWKT.eval(Constructors.scala:131)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection.apply(Unknown Source)
at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection.apply(Unknown Source)
at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)
at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)
at scala.collection.Iterator.foreach(Iterator.scala:941)
at scala.collection.Iterator.foreach$(Iterator.scala:941)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1429)
at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:162)
at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:160)
at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1429)
at scala.collection.TraversableOnce.aggregate(TraversableOnce.scala:219)
at scala.collection.TraversableOnce.aggregate$(TraversableOnce.scala:219)
at scala.collection.AbstractIterator.aggregate(Iterator.scala:1429)
at org.apache.spark.rdd.RDD.$anonfun$aggregate$2(RDD.scala:1207)
at org.apache.spark.SparkContext.$anonfun$runJob$6(SparkContext.scala:2290)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
21/09/18 11:51:43 INFO ShutdownHookManager: Shutdown hook called
21/09/18 11:51:43 INFO ShutdownHookManager: Deleting directory /tmp/spark-a846237b-773e-40e1-828b-76f646b1f247
21/09/18 11:51:43 INFO ShutdownHookManager: Deleting directory /tmp/spark-330d3500-8f92-463c-bd01-d27af27f4b71
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment