scala - 初始化 SparkContext : A master URL must be set in your configuration 时出错

标签 scala apache-spark k-means

我用过 this code

我的错误是:

Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties

17/02/03 20:39:24 INFO SparkContext: Running Spark version 2.1.0

17/02/03 20:39:25 WARN NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable

17/02/03 20:39:25 WARN SparkConf: Detected deprecated memory fraction 
settings: [spark.storage.memoryFraction]. As of Spark 1.6, execution and  
storage memory management are unified. All memory fractions used in the old 
model are now deprecated and no longer read. If you wish to use the old 
memory management, you may explicitly enable `spark.memory.useLegacyMode` 
(not recommended).

17/02/03 20:39:25 ERROR SparkContext: Error initializing SparkContext.

org.apache.spark.SparkException: A master URL must be set in your 
configuration
at org.apache.spark.SparkContext.<init>(SparkContext.scala:379)
at PCA$.main(PCA.scala:26)
at PCA.main(PCA.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.intellij.rt.execution.application.AppMain.main(AppMain.java:144)

17/02/03 20:39:25 INFO SparkContext: Successfully stopped SparkContext
Exception in thread "main" org.apache.spark.SparkException: A master URL must be set in your configuration
at org.apache.spark.SparkContext.<init>(SparkContext.scala:379)
at PCA$.main(PCA.scala:26)
at PCA.main(PCA.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
   sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at   
    sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

at java.lang.reflect.Method.invoke(Method.java:498)
at com.intellij.rt.execution.application.AppMain.main(AppMain.java:144)

Process finished with exit code 1

最佳答案

如果你单独运行 spark

val conf = new SparkConf().setMaster("spark://master") //missing 

并且您可以在提交作业时传递参数
spark-submit --master spark://master

如果你在本地运行 spark
val conf = new SparkConf().setMaster("local[2]") //missing 

您可以在提交作业时传递参数
spark-submit --master local

如果你在 yarn 上运行 Spark 那么
spark-submit --master yarn

关于scala - 初始化 SparkContext : A master URL must be set in your configuration 时出错,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/42032169/

相关文章:

scala - 为什么sortByKey的 Spark 这么慢?他们有其他选择吗?

python-3.x - Spark magic - 以字符串形式输入 sql 上下文

scala - 使用 Oozie 将 Spark 转为 Hbase

java - kmeans中的ArrayOutofBoundsException,在hadoop上运行时

scala - Play Framework 依赖注入(inject)对象与 @Singleton 类

scala - 如何映射扩展共同特征的对象列表?

linux - 使用 Play Framework 在 Linux 上本地进行 HTTPS

ruby - 如何使用 Ruby 进行一维 k 均值聚类?

python - 如何可视化用于 kmeans 聚类的 tf-idf 向量的数据点?

Scala:导入如何防止找到隐式值?