import pandas as pd
data = pd.read_csv('/Users/vishal/Desktop/Work/evdata.csv')
data = data[data.event == "left"]
data.head(3)
Date ID event
2017-04-25 firm_10 left
2017-01-18 firm_22 left
2017-09-08 firm_0 left
data['Date'] = pd.to_datetime(data['Date'])
data = data.groupby(['Date','ID']).agg({"event" : len}).unstack()
data.resample('B').sum()
DateID firm_0 firm_1 firm_10 firm_11
2017-01-09 0.0 0.0 0.0 0.0
2017-01-10 0.0 0.0 0.0 0.0
2017-01-11 0.0 0.0 1.0 0.0
Как это сделать в pyspark?