Skip to content
Advertisement

Tag: apache-spark

Read values from Java Map using Spark Column using java

I have tried below code to get Map values via spark column in java but getting null value expecting exact value from Map as per key search. and Spark Dataset contains one column and name is KEY and dataset name dataset1 values in dataset : Java Code – Current Output is: Expected Output : please me get this expected output.

why is my maven sub dependency version for spark connector package different from others

I am trying to use a pom file from a existing project and I am getting an error “Cannot resolve org.yaml:snakeyaml:1.15” What I find out about this error is that the com.datastax.spark:spark-cassandra-connector_2.11:2.5.0 uses a couple dependencies and a couple levels down it is using snakeyaml:1.15 which is quarantined by company proxy. Is there a way to specify for a given

Spark UDF function fail on Standalone Spark

I have spring boot java application myapp.jar with something udf function. SparkConfuration.java ToIntegerUdf.java sparkJars contains path to myJar.jar. Application build with Maven. Spark library version is 3.02 and scala version is 2.12.10. When I running application on Spark Standalone 3.0.2 I have an error: In spark worker log I see, worker fetch myJar: 21/03/23 19:33:24 INFO Executor: Fetching spark://demo.phoenixit.ru:39597/jars/myJar.jar with

Issue with Spark Big Query Connector with Java

Getting Below issue with the Spark Big Query connector in Dataproc cluster with below configuraton. Image: 1.5.21-debian10 Spark Version: 2.4.7 Scala Version: 2.12.10 This is working fine from local but failing when I deploy this in dataproc cluster.Can someone suggest some pointers for this issue? pom.xml: Here is the sample Code: Answer Can you please replace the Spark BigQuery connector

Advertisement