I'm seeing this too, but only on my personal cluster with the following config:
- Access Mode: Dedicated
- Policy: Unrestricted
- Runtime info: 15.4 LTS (includes Apache Spark 3.5.0, Scala 2.12)
If I use the IT department's shared clusters running either 14.3LTS or 16.2, both of which are running Access Mode = Shared, then the problem goes away.
Condition for the failure is odd.
- DF1 = pull some info from VIEW "foo"
- DF2= pull some different info from VIEW "foo"
- DF3=DF1.unionByName(DF2)
this would not seem to be a "self-join" as described in the error message, other than the fact that the first 2 dataframes are different cuts of the same view.