python - 每当我开始时,Pyspark都会给这些

标签 python hadoop apache-spark pyspark

每次我运行pyspark时,我都会得到这些错误,如果我在编写sc时忽略了这些错误,则会出现NameError:名称'sc'未定义任何帮助?

pyspark
Python 2.7.12 (default, Nov 19 2016, 06:48:10) 
[GCC 5.4.0 20160609] on linux2
Type "help", "copyright", "credits" or "license" for more information.
17/08/07 13:57:59 WARN NativeCodeLoader: Unable to load native-hadoop     library for your platform... using builtin-java classes where applicable
Traceback (most recent call last):
File "/usr/local/spark/python/pyspark/shell.py", line 45, in <module>
spark = SparkSession.builder\
File "/usr/local/spark/python/pyspark/sql/session.py", line 169, in getOrCreate
sc = SparkContext.getOrCreate(sparkConf)
File "/usr/local/spark/python/pyspark/context.py", line 334, in getOrCreate
SparkContext(conf=conf or SparkConf())
File "/usr/local/spark/python/pyspark/context.py", line 118, in __init__
conf, jsc, profiler_cls)
File "/usr/local/spark/python/pyspark/context.py", line 186, in _do_init
self._accumulatorServer = accumulators._start_update_server()
File "/usr/local/spark/python/pyspark/accumulators.py", line 259, in _start_update_server
server = AccumulatorServer(("localhost", 0), _UpdateRequestHandler)
File "/usr/lib/python2.7/SocketServer.py", line 417, in __init__
self.server_bind()
File "/usr/lib/python2.7/SocketServer.py", line 431, in server_bind
self.socket.bind(self.server_address)
File "/usr/lib/python2.7/socket.py", line 228, in meth
return getattr(self._sock,name)(*args)
socket.gaierror: [Errno -2] Name or service not known

最佳答案

经过1周的搜索,我刚刚找到解决方案,方法是将localhost添加到文件/ etc / hosts中,然后一切正常

关于python - 每当我开始时,Pyspark都会给这些,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/45546589/

相关文章:

apache-spark - 无法使用 Spark 2.2 访问 S3 数据

python - python中不可打印的字符串到数字

python - 如何将 object.__dict__ 再次转换为对象本身?

hadoop - 覆盖hadoop

scala - Spark分析异常: Path does not exist: Azure Blob Storage

hadoop - 如何在Hive中模拟BigQuery的分位数

python - 无法在 Pycharm 中创建或导入现有的 conda 项目

python - Pydub from_mp3 给出 [Errno 2] No such file or directory

hadoop - 每个文件的两个副本从本地复制到HDFS

hadoop - 无法更改或删除影响很大的分区表 - 由 : MetaException: Timeout when executing 引起