In Databricks Community Edition, the compute environment is set up as a single-node cluster. This means there is only one node, which serves both as the driver and the worker. Because of this, all data processingโincluding reading dataโis performed on this single node. There are no separate worker nodes available in the Community Edition, so you wonโt see distributed data processing across multiple workers. Thatโs why, in the Spark UI, it appears that all data is being read and processed on the driver node.
If you need true distributed processing with separate worker nodes, you would need to use the full (paid) version of Databricks or another Spark environment that supports multi-node clusters. In summary, the Community Edition does not support separate worker nodes; all operations happen on the single node available.
Aviral
AviralBhardwaj