โ01-18-2023 08:57 AM
I have a query that I'm trying to insert overwrite into a table. In an effort to try and speed up the query I added a range join hint. After adding it I started getting the error below.I can get around this though by creating a temporary view of the same query and caching that temporary view. It will then allow me to insert overwrite into the table.
โ01-18-2023 10:50 PM
Could you please check for the datatype of your source and target table.
There might be mismatch between both.
โ01-19-2023 05:57 AM
Both are the same
โ01-21-2025 06:59 AM
hi , can u help me in this
I am using this query to create a csv in a volume named test_volsrr that i created
INSERT OVERWRITE DIRECTORY '/Volumes/DATAMAX_DATABRICKS/staging/test_volsrr'
USING CSV OPTIONS ( 'delimiter' = ',' , 'header' = 'true' )
SELECT * FROM staging.extract1gb
DISTRIBUTE BY COALESCE( 1 );
i added DISTRIBUTE BY COALESCE(1) so that a single csv gets generated instead of multiple csvs , the size of extract1gb table is 1gb but the csv which is getting created is around 230gb , due to this it is taking more than an hour to execute . Can some pls explain this issue and a solution to generate the csv of optimal size so that execution becomes faster . I dont want to use pyspark .
โ01-30-2023 02:11 PM
Could you share your code and the full error stack trace please? Check the driver logs for the full stack trace.
Passionate about hosting events and connecting people? Help us grow a vibrant local communityโsign up today to get started!
Sign Up Now