- 10723 Views
- 7 replies
- 4 kudos
I am facing an error when trying to write data on Kafka using spark stream.#Extract
source_stream_df= (spark.readStream
.format("cosmos.oltp.changeFeed")
.option("spark.cosmos.container", PARM_CONTAINER_NAME)
.option("spark.cosmos.read.inferSchema.en...
- 10723 Views
- 7 replies
- 4 kudos
Latest Reply
What event hub namespace you were using?I had same problem and resolved by changing pricing plan from basic to standard as Kafka apps is not supporting in basic planLet me know if you had anything else. Thanks
6 More Replies
by
Sam
• New Contributor III
- 1471 Views
- 1 replies
- 4 kudos
Hello,I've noticed that Collect_Set and Collect_List are not pushed down to the database?Runtime DB 9.1LTSSpark 3.1.2Database: SnowflakeIs there any way to get a distinct set from a group by in a way that will push down the query to the database?
- 1471 Views
- 1 replies
- 4 kudos
Latest Reply
Hm so collect_set does not get translated to listagg.Can you try the following?use a more recent version of dbrxuse delta lake as spark sourceuse the latest version of the snowflake connectorcheck if pushdown to snowflake is enabled
- 1741 Views
- 0 replies
- 0 kudos
I used @pandas_udf write a function for speeding up the process(parsing xml file ) and then compare it's speed with single thread , Surprisingly , Using @pandas_udf is two times slower than single-thread code. And the number of xml files I need to p...
- 1741 Views
- 0 replies
- 0 kudos