Я использую spark 2.4.0 в Google Cloud Compute Engine с CentOS 6 и 3,75 ГМ памяти.Когда я пытался перейти в режим pyspark, он выдает следующую ошибку:
[user@pb-instance bin]$ pyspark
Python 2.6.6 (r266:84292, Aug 18 2016, 15:13:37)
[GCC 4.4.7 20120313 (Red Hat 4.4.7-17)] on linux2
Type "help", "copyright", "credits" or "license" for more information.
Traceback (most recent call last):
File "/home/user/spark/spark-2.4.0-bin-hadoop2.7/python/pyspark/shell.py", line 31, in <module>
from pyspark import SparkConf
File "/home/user/spark/spark-2.4.0-bin-hadoop2.7/python/pyspark/__init__.py", line 51, in <module>
from pyspark.context import SparkContext
File "/home/user/spark/spark-2.4.0-bin-hadoop2.7/python/pyspark/context.py", line 31, in <module>
from pyspark import accumulators
File "/home/user/spark/spark-2.4.0-bin-hadoop2.7/python/pyspark/accumulators.py", line 97, in <module>
from pyspark.serializers import read_int, PickleSerializer
File "/home/user/spark/spark-2.4.0-bin-hadoop2.7/python/pyspark/serializers.py", line 71, in <module>
from pyspark import cloudpickle
File "/home/user/spark/spark-2.4.0-bin-hadoop2.7/python/pyspark/cloudpickle.py", line 246, in <module>
class CloudPickler(Pickler):
File "/home/user/spark/spark-2.4.0-bin-hadoop2.7/python/pyspark/cloudpickle.py", line 270, in CloudPickler
dispatch[memoryview] = save_memoryview
NameError: name 'memoryview' is not defined
>>>