apache-spark - Spark 提交给YARN时出现 “unknown queue: root.default”错误

标签 apache-spark hadoop pyspark airflow yarn

我正在通过Airflow和SparkSubmitOperator向新建的YARN集群提交一个简单的Pyspark单词计数作业。作业达到了YARN,我可以在ResourceManager UI中看到它,但是由于以下错误而失败:

“诊断:用户root向未知队列提交的应用程序application_1582063076991_0002:root.default”

*User:  root
Name:   PySpark Wordcount
Application Type:   SPARK
Application Tags:   
YarnApplicationState:   FAILED
Queue:  root.default
FinalStatus Reported by AM: FAILED
Started:    Fri Feb 21 08:01:25 +1100 2020
Elapsed:    0sec
Tracking URL:   History
Diagnostics:    Application application_1582063076991_0002 submitted by user root to unknown queue: root.default*

default.root队列似乎确实存在:
*Application Queues
  Legend:CapacityUsedUsed (over capacity)Max Capacity
      .root    0.0% used
          ..Queue: default    0.0% used


'default' Queue Status
Queue State:    RUNNING
Used Capacity:  0.0%
Configured Capacity:    100.0%
Configured Max Capacity:    100.0%
Absolute Used Capacity: 0.0%
Absolute Configured Capacity:   100.0%
Absolute Configured Max Capacity:   100.0%
Used Resources: <memory:0, vCores:0>
Num Schedulable Applications:   0
Num Non-Schedulable Applications:   0
Num Containers: 0
Max Applications:   10000
Max Applications Per User:  10000
Max Application Master Resources:   <memory:3072, vCores:1>
Used Application Master Resources:  <memory:0, vCores:0>
Max Application Master Resources Per User:  <memory:3072, vCores:1>
Configured Minimum User Limit Percent:  100%
Configured User Limit Factor:   1.0
Accessible Node Labels: *
Preemption: disabled*

我在这里想念什么?谢谢

最佳答案

用队列名称default提交。

资源管理器中的root仅用于以分层形式对队列进行分组。

关于apache-spark - Spark 提交给YARN时出现 “unknown queue: root.default”错误,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/60328948/

相关文章:

python - pyspark Dataframe API cast ('timestamp' ) 不适用于时间戳字符串

hadoop - MapReduce寻找社交网络影响力

hadoop - 如何垂直而不是水平拆分数据?

apache-spark - 间隔 30 天转换为间隔 4 周 2 天

python - 如何通过pyspark读取gz压缩文件

hadoop - 在lib文件夹中找不到与hadoop-2.7.3相关的jar

python - pickle Spark RDD 并将其读入 Python

hadoop - Spark的API newHadoopRDD到底做什么?

apache-spark - Apache spark 主服务器未启动。由 : java. lang.reflect.InaccessibleObjectException 引起

python - 在 Spark 数据框中为用户的每个项目创建每一行