hadoop - 为什么在Logistic回归期间将SparkContext关闭?

标签 hadoop apache-spark

我认为它与内存有关,因为它适用于较小的数据集。该程序在使用Spark-Mllib的Logistic回归的同时,会过早地关闭它。我在下面运行此命令以在HDFS上启动我的Spark程序。

export SPARK_CONF_DIR=/home/gs/conf/spark/latest
export SPARK_HOME=/home/gs/spark/latest

$SPARK_HOME/bin/spark-submit --class algoRunner --master yarn --deploy-mode cluster --conf spark.dynamicAllocation.enabled=true \
 --executor-memory 8g  --queue default --conf spark.hadoop.hive.querylog.location='${java.io.tmpdir}/hivelogs' \
~/spark/Product-Classifier-Pipeline-assembly-1.0.jar

我收到以下错误:
17/08/02 21:53:40 ERROR ApplicationMaster: RECEIVED SIGNAL TERM
17/08/02 21:53:40 INFO SparkContext: Invoking stop() from shutdown hook
17/08/02 21:53:40 INFO SparkUI: Stopped Spark web UI at http://gsrd219n01.red.ygrid.yahoo.com:45546
17/08/02 21:53:40 INFO DAGScheduler: Job 10 failed: treeAggregate at LogisticRegression.scala:1670, took 2.351935 s
17/08/02 21:53:40 INFO DAGScheduler: ShuffleMapStage 19 (treeAggregate at LogisticRegression.scala:1670) failed in 1.947 s due to Stage cancelled because SparkContext was shut down
17/08/02 21:53:40 ERROR LiveListenerBus: SparkListenerBus has already stopped! Dropping event SparkListenerStageCompleted(org.apache.spark.scheduler.StageInfo@21bec75d)
17/08/02 21:53:40 ERROR LiveListenerBus: SparkListenerBus has already stopped! Dropping event SparkListenerJobEnd(10,1501710820713,JobFailed(org.apache.spark.SparkException: Job 10 cancelled because SparkContext was shut down))
17/08/02 21:53:40 ERROR ApplicationMaster: User class threw exception: org.apache.spark.SparkException: Job 10 cancelled because SparkContext was shut down
org.apache.spark.SparkException: Job 10 cancelled because SparkContext was shut down

最佳答案

驱动程序内存不足。增加它可以防止这些错误。

关于hadoop - 为什么在Logistic回归期间将SparkContext关闭?,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/45471790/

相关文章:

apache-spark - 更新和插入

hadoop - SequenceFile替代/扩展,允许就地更新

hadoop - oozie 中的 org.apache.hadoop.mapred.lib.MultipleOutputs.addNamedOutput()

scala - 如何优雅地提前结束 Spark 应用程序?

java - Spark 给出错误 : Files\Java\jdk1. 8.0_13 1""= ="x"此时意外

java - 使用 Apache Spark 和 Java 增量加载 parquet 文件中的数据

python - 在 Spark 中广播 Annoy 对象(对于最近的邻居)?

hadoop - MapReduceIndexerTool输出目录错误 "Cannot write parent of file"

hadoop - 是否可以使用多个 jar 来指定 hadoop 中的作业?

hadoop - hadoop mapreduce程序中的错误