I can\'t figure it out, but guess it\'s simple. I have a spark dataframe df. This df has columns \"A\",\"B\" and \"C\". Now let\'s say I have an Array containing the name of
Since Spark 1.4 you can use drop method:
Scala:
case class Point(x: Int, y: Int) val df = sqlContext.createDataFrame(Point(0, 0) :: Point(1, 2) :: Nil) df.drop("y")
Python:
df = sc.parallelize([(0, 0), (1, 2)]).toDF(["x", "y"]) df.drop("y") ## DataFrame[x: bigint]