PySpark: How to covert column with Ljava.lang.Object
问题 I created data frame in PySpark by reading data from HDFS like this: df = spark.read.parquet('path/to/parquet') I expect the data frame to have two column of strings: +------------+------------------+ |my_column |my_other_column | +------------+------------------+ |my_string_1 |my_other_string_1 | |my_string_2 |my_other_string_2 | |my_string_3 |my_other_string_3 | |my_string_4 |my_other_string_4 | |my_string_5 |my_other_string_5 | |my_string_6 |my_other_string_6 | |my_string_7 |my_other