@Danial Malik :
To connect Databricks running on Azure with RDS running on AWS, you can use JDBC or ODBC connectivity.
Here are the general steps:
- Set up a Virtual Private Network (VPN) or a peering connection between your Azure and AWS environments to ensure secure communication between them.
- Install the JDBC or ODBC driver for RDS on the Databricks cluster. You can find the driver download and installation instructions in the AWS RDS documentation.
- Configure the JDBC or ODBC connection to connect to the RDS database using the appropriate driver, database URL, username, and password. You may also need to configure additional parameters, such as SSL settings, depending on your RDS configuration.
- Write Spark code in Databricks to read data from RDS using the JDBC or ODBC connection, and then write the data to Azure DataLake.
Regarding the additional layer of basic http authentication on RDS, you will need to provide the appropriate credentials in the JDBC or ODBC connection configuration.
Keep in mind that cross-cloud data transfer can have performance and security implications, so you may want to consider optimizing your network and data transfer protocols to ensure efficient and secure data transfer.
Hope this helps!