Often times, data is available with a folder structure like,
2000-01-01/john/smith
rather than the Hive partition spec,
date=2000-01-01/first_name=john/last_name=smith
Spark (and pyspark) can read partitioned data easily when using the Hive folder structure, but with the "bad" folder structure it becomes difficult and involved regex and things.
Is there an easier way to deal with non-hive folder structure for partitioned data in Spark?