0
from pyspark.sql import *
from pyspark import SQLContext
sqc=SQLContext(sc)
input=sc.textFile("file:///home/cloudera/Desktop/uber.txt")
df=input.map(lambda x:x.split(","))
df=sqc.createDataFrame(input.map(lambda x:x.split(","))
input.map(lambda r:Row(basedid=r[0],dt=r[1],nveh=int(r[2]),ncus=int(r[3])))))
當我執行上面的代碼時,我得到以下錯誤。pyspark管道錯誤代碼下面的代碼
TypeError: 'PipelinedRDD' object is not callable