in pandas , I can do something like this .
data = {"col1" : [np.random.randint(10) for x in range(1,10)],
"col2" : [np.random.randint(100) for x in range(1,10)]}
mypd = pd.DataFrame(data)
mypd
and give the two columns
are there any similar way to create a spark dataframe in pyspark ?
