โ01-27-2025 03:10 AM
I followed the official Databricks documentation("https://docs.databricks.com/en/_extras/notebooks/source/mongodb.html")
to integrate MongoDB Atlas with Spark by setting up the MongoDB Spark Connector and configuring the connection string in my Databricks cluster. However, I am encountering issues when trying to read data from MongoDB using Spark.
While I can successfully connect to MongoDB using the MongoClient in Python and execute queries like
I am unable to load data using the Spark connector with the following code:
The connection string is the same in both cases, and I have confirmed that the necessary permissions and IP whitelisting are correctly configured in MongoDB Atlas.
Despite this, no data is being retrieved when using Spark, and Iโm unable to identify the issue.
also, I attached error screenshot below.
Can anyone provide guidance on potential configuration issues or additional steps needed to troubleshoot this problem with the MongoDB Spark connector in Databricks?
โ01-27-2025 04:34 AM
Hi @vidya_kothavale ,
Could you try to change "spark.mongodb.input.uri" to following?
spark.read.format("mongodb").option("spark.mongodb.read.connection.uri"
โ01-27-2025 04:34 AM
Hi @vidya_kothavale ,
Could you try to change "spark.mongodb.input.uri" to following?
spark.read.format("mongodb").option("spark.mongodb.read.connection.uri"
โ01-27-2025 05:38 AM
Thanks! @szymon_dybczak It's working perfectly now.
โ03-01-2025 05:45 AM - edited โ03-01-2025 05:46 AM
Hi vidya. I have the same problem.
I can connect using pymongo and compass . I installed the library org.mongodb.spark:mongo-spark-connector_2.13:10.4.1 (latest one) on my cluster using the runtime 16.2 but I never was able to connect to same mongo cluster (sharded) using the primary as default.
This is the scala code (I've tested in python as well)
Passionate about hosting events and connecting people? Help us grow a vibrant local communityโsign up today to get started!
Sign Up Now