first question here, so I apologise if something isn't clear. I am new to pySpark, and using databricks I was trying to read in an excel file saved as a csv with the following code
df = spark.read.csv('/FileStore/tables/file.csv',
sep = ";",
inferSchema = "true",
header = "true")
This works fine, except some of the observations get null values, while in the excel file there are no null values. The actual values can be found in other rows. Maybe better explained with an example: If the excel file has the row A B C D Then it becomes in the table (for some rows):
A B null null
C D null null
My question is how could I fix this? Thanks in advance