by
juanc
• New Contributor II
- 2616 Views
- 9 replies
- 2 kudos
It would be possible to activate a custom extensions like Sedona (https://sedona.apache.org/download/databricks/ ) in SQL Endopoints?Example error:java.lang.ClassNotFoundException: org.apache.spark.sql.sedona_sql.UDT.GeometryUDT at org.apache.spark....
- 2616 Views
- 9 replies
- 2 kudos
Latest Reply
@Kaniz What is the right way to add custom spark extension to sql warehouse clusters?
8 More Replies
- 1123 Views
- 2 replies
- 1 kudos
The usual way to use Apache Sedona inside pySpark is by first registering Sedona types and functions withSedonaRegistrator.registerAll(spark)We need to have these autoregistered when the cluster start (to be able, for example, to perform geospatial q...
- 1123 Views
- 2 replies
- 1 kudos
Latest Reply
Hi @Giovanni Allegri​ Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answ...
1 More Replies
- 1927 Views
- 1 replies
- 1 kudos
Hello,I know how to create .shp file from Geopandas dataframe using code similar to this, also mentioned on SO:gpd_df = geopandas.GeoDataFrame(pandas_df, geometry='geom')
gpd_df .to_file("username/nh.shp")However I have .parquet files that I can load...
- 1927 Views
- 1 replies
- 1 kudos
Latest Reply
@Bartosz Maciejewski​ :Spark does not have native support for writing Shapefiles directly. However, you can use a third-party library such as GeoPandas or PyShp to write your Spark DataFrame to a Shapefile.Here's an example of how to use GeoPandas to...