Я не могу понять, почему я получаю эту конкретную ошибку при запуске моего python файла:
def main():
# Creating Spark context object
sc = SparkContext(appName="Fake Opinion Detector")
spark = SparkSession(sc)
# Loading dataset files
meta_data = sc.textFile(sys.argv[1]).map(lambda x: x.split("\t")).toDF(["user_id","prod_id","rating","label","date"])
review_data = sc.textFile(sys.argv[2]).map(lambda x: x.split("\t")).toDF(["user_id","prod_id","date","review"])
Ошибка, которую я получаю:
20/04/04 14:09:30 WARN Utils: Your hostname, user-Lenovo-G50-70 resolves to a loopback address: 127.0.1.1; using 192.168.43.58 instead (on interface wlx00e04d0cbda4)
20/04/04 14:09:30 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address
20/04/04 14:09:31 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
Traceback (most recent call last):
File "FakeOpinionDetection.py", line 369, in <module>
main()
File "FakeOpinionDetection.py", line 321, in main
meta_data = sc.textFile(sys.argv[1]).map(lambda x: x.split("\t")).toDF(["user_id","prod_id","rating","label","date"])
IndexError: list index out of range