While writing parquet file back to DataLake Gen2 is creating additional files.
Example:
%python
rawfile = "wasbs://[email protected]/xxxx/2019-09-30/account.parquet"
curatedfile = "wasbs://[email protected]/xxxx-Curated/2019-09-30/account.parquet"
dfraw = spark.read.parquet(rawfile)
dfraw.write.parquet(curatedfile, mode = "overwrite")
display(dfraw)
File name supplied (account.parquet) is accounted to created folder rather creating file with that name.
How can these additional files be ignored and file written with name supplied.