我在Windows上安裝了Spark,但無法啓動pyspark
。當我輸入c:\Spark\bin\pyspark
時,出現以下錯誤:無法運行pyspark
Python 3.6.0 |Anaconda custom (64-bit)| (default, Dec 23 2016, 11:57:41) [MSC v.1900 64 bit (AMD64)] on win32 Type "help", "copyright", "credits" or "license" for more information. Traceback (most recent call last): File "c:\Spark\bin..\python\pyspark\shell.py", line 30, in import pyspark File "c:\Spark\python\pyspark__init__.py", line 44, in from pyspark.context import SparkContext File "c:\Spark\python\pyspark\context.py", line 36, in from pyspark.java_gateway import launch_gateway File "c:\Spark\python\pyspark\java_gateway.py", line 31, in from py4j.java_gateway import java_import, JavaGateway, GatewayClient File "", line 961, in _find_and_load File "", line 950, in _find_and_load_unlocked File "", line 646, in _load_unlocked File "", line 616, in _load_backward_compatible File "c:\Spark\python\lib\py4j-0.10.4-src.zip\py4j\java_gateway.py", line 18, in File "C:\Users\Eigenaar\Anaconda3\lib\pydoc.py", line 62, in import pkgutil File "C:\Users\Eigenaar\Anaconda3\lib\pkgutil.py", line 22, in ModuleInfo = namedtuple('ModuleInfo', 'module_finder name ispkg') File "c:\Spark\python\pyspark\serializers.py", line 393, in namedtuple cls = _old_namedtuple(*args, **kwargs) TypeError: namedtuple() missing 3 required keyword-only arguments: 'verbose', 'rename', and 'module'
我在做什麼錯在這裏?
我能說出幾件事情,你做錯了 - _(1)_你格式化你的日誌條目爲「塊引用」(用於引用純文本文檔,如)而不是「代碼樣本」; _(2)_你沒有提及你已經安裝的Spark版本:_(3)_你甚至沒有嘗試谷歌spark類型錯誤:namedtuple()缺少3個必需的關鍵字參數, //issues.apache.org/jira/browse/SPARK-19019作爲最佳答案。 –
請儘量在回溯中保留行結束,以使其遠程可讀。 – Eric