You can now read data from another #databricks workspace using a native JDBC driver with the "spark.read.format("databricks")" or "CREATE TABLE databricks_external_table USING" databricks commands. While delta sharing is typically preferred, this new feature makes it simple for example, to transfer data between regions where delta sharing may not be an option.