cancel
Showing results for 
Search instead for 
Did you mean: 
Nagarathna
New Contributor II
since ‎04-15-2024
‎05-19-2025

User Stats

  • 4 Posts
  • 0 Solutions
  • 0 Kudos given
  • 0 Kudos received

User Activity

Hi team,I have a dataframe with 1269408570800 rows . I need to write this data to unity catalog table.How can I upload huge quantity of data ?I'm using databricks i runtime 15.4 LTS with 4 workers and each worker type is i3.4xlarge and driver of type...
I am trying to reading json from aws s3 using with open in databricks notebook using shared cluster.Error message:No such file or directory:'/dbfs/mnt/datalake/input_json_schema.json'In single instance cluster the above error is not found.