I am trying to work through Tutorial: Query data from a Notebook.
Access errors are defeating my attempts. Steps to reproduce:
- sign up for free trial through Databricks website. The path skipped the subscription-selection step and defaulted the trial to Premium.
- create workspace using AWS Quick Start
- create a cluster. Runtime 13.3
- create a notebook
- step 3 option 1, Create Spark table from the CSV data
- expected behavior: table created successfully
- actual behavior: error UnityCatalogServiceException: [RequestId=d548e435-f8ec-49e4-b134-520d96b27cd6 ErrorClass=INVALID_PARAMETER_VALUE] GenerateTemporaryPathCredential uri databricks-datasets/Rdatasets/data-001/csv/ggplot2/diamonds.csv is not a valid URI. Error message: INVALID_PARAMETER_VALUE: Missing cloud file system scheme.
- Step 3 option 2, Write the CSV data to Delta Lake format and create a Delta table
- expected behavior: both steps complete successfully
- actual behavior: step 1 completes successfully. Step 2 ("Create a Delta table at the stored location") fails with error UnityCatalogServiceException: [RequestId=89804dc3-8c28-43f0-865b-aa80e0a1bd0d ErrorClass=INVALID_PARAMETER_VALUE] GenerateTemporaryPathCredential uri /mnt/delta/diamonds is not a valid URI. Error message: INVALID_PARAMETER_VALUE: Missing cloud file system scheme.
Reading past forum messages suggests that the issue results from lack of storage credentials and/or external locations (https://docs.databricks.com/en/data-governance/unity-catalog/manage-external-locations-and-credentia...). This seems like an issue that the stack created through the AWS Quick Start setup would have dealt with. Is there an incompatibility between the tutorial instructions and use of Unity Catalog?
Can anyone suggest how to resolve this issue with the least effort?