- 2968 Views
- 3 replies
- 1 kudos
The usual way to use Apache Sedona inside pySpark is by first registering Sedona types and functions withSedonaRegistrator.registerAll(spark)We need to have these autoregistered when the cluster start (to be able, for example, to perform geospatial q...
- 2968 Views
- 3 replies
- 1 kudos
Latest Reply
Hi @Giovanni Allegri Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answ...
2 More Replies
by
juanc
• New Contributor II
- 5991 Views
- 8 replies
- 2 kudos
It would be possible to activate a custom extensions like Sedona (https://sedona.apache.org/download/databricks/ ) in SQL Endopoints?Example error:java.lang.ClassNotFoundException: org.apache.spark.sql.sedona_sql.UDT.GeometryUDT at org.apache.spark....
- 5991 Views
- 8 replies
- 2 kudos
Latest Reply
@Retired_mod What is the right way to add custom spark extension to sql warehouse clusters?
7 More Replies
- 3731 Views
- 1 replies
- 1 kudos
Hello,I know how to create .shp file from Geopandas dataframe using code similar to this, also mentioned on SO:gpd_df = geopandas.GeoDataFrame(pandas_df, geometry='geom')
gpd_df .to_file("username/nh.shp")However I have .parquet files that I can load...
- 3731 Views
- 1 replies
- 1 kudos
Latest Reply
@Bartosz Maciejewski :Spark does not have native support for writing Shapefiles directly. However, you can use a third-party library such as GeoPandas or PyShp to write your Spark DataFrame to a Shapefile.Here's an example of how to use GeoPandas to...