-
Using Scala version 2.12.14 (OpenJDK 64-Bit Server VM, Java 1.8.0_322) scala> import org.apache.doris.spark._ scala> val dorisSparkRDD = sc.dorisRDD( scala> dorisSparkRDD.collect()` 22/05/10 10:30:46 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 0.0 (TID 2) (cluster-6a52-w-0.us-central1-a.c.xed-project-237404.internal executor 1): java.lang.ClassNotFoundException: org.apache.doris.spark.rdd.DorisPartition currently env: the above code works fine in standalone mode. the connector was built in docker support by official ( centos) 1.0.0-rc03 thanks in advance for any clue ~ |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
after help from community, this article the package folder which located according to config file : $SPARK_HOME/conf/spark-defaults.conf by appending the hdfs file path , and retry will work
|
Beta Was this translation helpful? Give feedback.
after help from community, this article
Yarn集群模式运行的Spark,则将此文件放入预部署包中。
预部署包 or so called pre-deployment package
the package folder which located according to config file : $SPARK_HOME/conf/spark-defaults.conf
spark.yarn.jars=local:/usr/lib/spark/jars/*
by appending the hdfs file path , and retry will work
spark.yarn.jars=local:/usr/lib/spark/jars/*,hdfs:///spark-jars/doris-spark-connector-3.1.2-2.12-1.0.0.jar
#