06-23-2022 02:59 AM
Currently I am having some issues with the writing of the parquet file in the Storage Container. I do have the codes running but whenever the dataframe writer puts the parquet to the blob storage instead of the parquet file type, it is created as a folder type with many files content to it.
One note to it, I tried searching various ways in the internets that it seems this is default creation when using pyspark and I can see in the folder created there was a file parquet with a snappy add into it (refer to the screenshots below)
If this is a default creation to pyspark code, how can I write a single parquet format that will do some splitting or creation of folder? any recommendations? on how to do it?
06-24-2022 06:16 AM
When you write a file, it uses the default compression if you don't specify it. The default compression is snappy, so that's expected + desired behavior.
Parquet is meant to be splittable. It also needs to create the other files that begin with the underscore to ensure you don't get partial or broken writes.
What exactly are you trying to do?
06-24-2022 06:16 AM
When you write a file, it uses the default compression if you don't specify it. The default compression is snappy, so that's expected + desired behavior.
Parquet is meant to be splittable. It also needs to create the other files that begin with the underscore to ensure you don't get partial or broken writes.
What exactly are you trying to do?
06-27-2022 07:02 AM
Already found out that this is already the behaviour, so to make it work, currently all the wrangled folders are being deleted and the file parquet contents inside the wrangled folder are already being moved outside the folder while renaming it. That's the solution I see from my goal to dump a single parquet file on the container with no wrangled folders.
Thank you @Joseph Kambourakis
06-27-2022 04:29 AM
Hello @Karl Saycon
Can you try setting this config to prevent additional parquet summary and metadata files from being written? The result from dataframe write to storage should be a single file.
A combination of below three properties will help to disable writing all the transactional files which start with "_".
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group