cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Any on please suggest how we can effectively loop through PySpark Dataframe .

Ancil
Contributor II

Scenario: I Have a dataframe with more than 1000 rows, each row having a file path and result data column. I need to loop through each row and write files to the file path, with data from the result column.

what is the easiest and time effective way to do this?

I tried with collect and it's taking long time.

And I tried UDF methods but getting below error

image

11 REPLIES 11

UmaMahesh1
Honored Contributor III

Hi @Ancil P A​ 

Is your data in the result column a json value or how is it ?

From your question, I understood that you have two columns in your df, 1 column is the file path and the other column is data.

Also please post what udf you are trying to build so that if your approach is useful, fix can be done on that.

Cheers..

Uma Mahesh D

Hi @Uma Maheswara Rao Desula​ 

In the result column have result json data , but column type is string.

Please find below screen shot for UDF

imageOnce I called below line am getting below error

input_data_df = input_data_df.withColumn("is_file_created",write_files_udf(col("file_path"),col("data_after_grammar_correction")))

image

-werners-
Esteemed Contributor III

Is it an option to write is as a single parquet file, but partitioned?

Like that physically the paths of the partitions are different, but they all belong to the same parquet file.

The key is to avoid loops.

Hi @Werner Stinckens​ 

My use case is to write text files with how many rows in dataframe.

For example, if I have 100 rows, then I need to write 100 files in the specified location.

-werners-
Esteemed Contributor III

yes exactly, that is what partitioning does.

all you need is a common path where you will write all those files, and partition on the part that is not common.

f.e.

/path/to/file1|<data>

/path/to/file2|<data>

the common part(/path/to), you use as target location.

The changing part (file1, file2) you use as partition column

so it will become:

df.write.partitionBy(<fileCol>).parquet(<commonPath>)

Spark will write a file (or even more than 1) per partition.

If you want only one single file you also have to repartition by filecol.

Ancil
Contributor II

Hi @Werner Stinckens​ 

In my case there is no common path, the file path column has different paths in a storage container.

Do we have any other way

-werners-
Esteemed Contributor III

afaik partitioning is the only way to write to multiple locations in parallel.

This SO thread perhaps has a way.

Ancil
Contributor II

Thanks a lot, let me check

Ancil
Contributor II

Hi @Werner Stinckens​ 

After partitioning also am getting below error. Do u have any about this error

image

NhatHoang
Valued Contributor II

Hi,

I agree with Werners, try to avoid loop with Pyspark Dataframe.

If your dataframe is small, as you said, only about 1000 rows, you may consider to use Pandas.

Thanks.​

Hi @Nhat Hoang​ 

The size may vary it may be up to 1 lakh, I will check with pandas

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group