- 4839 Views
- 3 replies
- 1 kudos
py4j.security.Py4JSecurityException: Method public com.databricks.backend.daemon.dbutils.DBUtilsCore$Result com.databricks.backend.daemon.dbutils.DBUtilsCore.mount(java.lang.String,java.lang.String,java.lang.String,java.lang.String,java.util.Map) is ...
- 4839 Views
- 3 replies
- 1 kudos
Latest Reply
Hi , have you sorted this issue ? can you please let me know the solution?
2 More Replies
by
ACK
• New Contributor II
- 3520 Views
- 2 replies
- 2 kudos
Hi,I have a method named main it takes **kwargs as a parameter. def main(**kwargs):
parameterOne = kwargs["param-one"]
parameterTwo = kwargs["param-two"]
parameterThree = kwargs["param-optional-one"] if "param-optional-one" in kwargs else...
- 3520 Views
- 2 replies
- 2 kudos
Latest Reply
it is command-line parameters so it is like ---param-one=testyou can test it with ArgumentParserfrom argparse import ArgumentParser
parser = ArgumentParser()
parser.add_argument("--param-one", dest="parameterOne")
args = parser.parse_args()
1 More Replies
- 7687 Views
- 0 replies
- 0 kudos
Hi everyone, I am having some troubles to add a deduplication step on a file streaming that is already running. The code I am trying to add is this one:df = df.withWatermark("arrival_time", "20 minutes")\
.dropDuplicates(["event_id", "arrival_time"])...
- 7687 Views
- 0 replies
- 0 kudos
- 2088 Views
- 3 replies
- 2 kudos
I already have a trained and saved model that was created outside of MLflow. What is the best way to handle it if I want this model to be added to an MLflow experiment?
- 2088 Views
- 3 replies
- 2 kudos
Latest Reply
Hi @Trevor Bishop Just wanted to check in if you were able to resolve your issue or do you need more help? We'd love to hear from you.Thanks!
2 More Replies
- 4357 Views
- 4 replies
- 1 kudos
- 4357 Views
- 4 replies
- 1 kudos
Latest Reply
I think it is because you did not attach the libraries to the cluster.When you work with a notebook, the sparksession is already created.To add libraries, you should install them on the cluster (in the compute tab) using f.e. pypi/maven etc.
3 More Replies
by
wyzer
• Contributor II
- 2351 Views
- 2 replies
- 1 kudos
Hello,We are new on Databricks and we would like to know if our working method are good.Currently, we are working like this :spark.sql("CREATE TABLE Temp (SELECT avg(***), sum(***) FROM aaa LEFT JOIN bbb WHERE *** >= ***)")With this method, are we us...
- 2351 Views
- 2 replies
- 1 kudos
Latest Reply
Spark will handle the map/reduce for you.So as long as you use Spark provided functions, be it in scala, python or sql (or even R) you will be using distributed processing.You just care about what you want as a result.And afterwards when you are more...
1 More Replies
by
missyT
• New Contributor III
- 1710 Views
- 1 replies
- 3 kudos
I do a lot of work with numpy arrays and pytorch tensors, but occasionally throw some native lists around. I naturally want to write <list>.sum(), which would work for these other third-party iterables, but doesn't work for native lists.It'd be very ...
- 1710 Views
- 1 replies
- 3 kudos
Latest Reply
I think reason is that list can contain different type of objects than just integers and floats (so nested lists, string and all possible other kind of objects) so it doesn't make sense to implement .sum method as it would fail in many cases.
- 2334 Views
- 0 replies
- 0 kudos
The repos API has a patch method to update a repo in the workspace (to do a git pull).We would please like to verify: is this method fully synchronous? Is it guaranteed to only return a 200 after the update is complete? Or, would immediately referenc...
- 2334 Views
- 0 replies
- 0 kudos