- 1391 Views
- 1 replies
- 0 kudos
Okta Integration
My company uses Okta as a SSO provider. Can I integrate Okta with Databricks for a SSO experience?
- 1391 Views
- 1 replies
- 0 kudos
- 0 kudos
Yes, okta is among the supported identity providers. Read more here : https://docs.databricks.com/administration-guide/users-groups/single-sign-on/index.html
- 0 kudos
- 1188 Views
- 1 replies
- 0 kudos
- 1188 Views
- 1 replies
- 0 kudos
- 0 kudos
Yes. There is a property called dbus_per_hour that you can add to your cluster policy.See https://docs.databricks.com/administration-guide/clusters/policies.html#cluster-policy-virtual-attribute-pathsHere's an example policy that uses it:https://docs...
- 0 kudos
- 7154 Views
- 1 replies
- 0 kudos
Resolved! My cluster is running an init script, and I want to see what's going on.
After I log in to the workspace, where can I find the logs?
- 7154 Views
- 1 replies
- 0 kudos
- 0 kudos
See public docs: https://docs.databricks.com/clusters/init-scripts.html#cluster-scoped-init-script-logsDon't forget to enable cluster log delivery: https://docs.databricks.com/clusters/configure.html#cluster-log-deliveryNote that this only works for ...
- 0 kudos
- 2797 Views
- 1 replies
- 0 kudos
- 2797 Views
- 1 replies
- 0 kudos
- 0 kudos
Usually, your account URL is where you navigate to log in.
- 0 kudos
- 2251 Views
- 1 replies
- 0 kudos
- 2251 Views
- 1 replies
- 0 kudos
- 0 kudos
At a high level a Lakehouse must contain the following properties:Open direct access data formats (Apache Parquet, Delta Lake etc.)First class support for machine learning and data science workloadsstate of the art performance Databricks is the firs...
- 0 kudos
- 3197 Views
- 1 replies
- 0 kudos
Extending the duration of Ganglia metrics logs
Any insights on how to analyse Ganglia Metrics logs for an extended duration of time, not just 15 minute snapshots? We need to visualize cluster CPU utilization for the duration of cluster uptime
- 3197 Views
- 1 replies
- 0 kudos
- 0 kudos
One option here would be integration with Observability tools such as Datadog which can capture the cluster metrics on a more NRT basis. More details are here - https://docs.datadoghq.com/integrations/databricks/?tab=driveronly
- 0 kudos
- 1720 Views
- 0 replies
- 0 kudos
What's the right batch size in deep learning training?
Using Ganglia you can monitor how busy is the GPU(s). Increasing the batch size would increase that utilization. Bigger batches improve how well each batch updates the model (up to a point) with more accurate gradients. That in turn can allow trainin...
- 1720 Views
- 0 replies
- 0 kudos
- 4529 Views
- 0 replies
- 0 kudos
What's Early Stopping in Hyperopt? When should it be used?
It’s advantageous to stop running trials if progress has stopped. Hyperopt offers an early_stop_fn parameter, which specifies a function that decides when to stop trials before max_evals has been reached. Hyperopt provides a function no_progress...
- 4529 Views
- 0 replies
- 0 kudos
- 2004 Views
- 1 replies
- 0 kudos
Resolved! How to determine if am using the same DBR minor version?
DBR minor version details are not exposed. However, in the documentation, it mentioned Databricks performs maintenance releases every 2 weeks. How can I determine if I am using the same minor version
- 2004 Views
- 1 replies
- 0 kudos
- 0 kudos
The below code snippet can help to determine the DBR Hash string for the DBR version. DBR hash string is unique for the DBR minor version. val scalaVersion = scala.util.Properties.versionString val hadoopVersion = org.apache.hadoop.util.VersionInf...
- 0 kudos
- 2355 Views
- 2 replies
- 0 kudos
- 2355 Views
- 2 replies
- 0 kudos
- 0 kudos
Delete workspace doesn't delete the root bucket. You could choose to use the same root bucket for more than one workspace ( though not recommended ) It is recommended to automate the infrastructure creation via terraform or quickstart so that cleanup...
- 0 kudos
- 3177 Views
- 1 replies
- 0 kudos
Monitoring jobs
Are there any event streams that are or could be exposed in AWS (such as Cloudwatch Eventbridge events or SNS messages? In particular I'm interested in events that detail jobs being run. The use case here would be for monitoring jobs from our web app...
- 3177 Views
- 1 replies
- 0 kudos
- 0 kudos
You could write code to call the PutLogEvents api at the beginning of each job to write out custom events to cloudwatch / or use aws sdk to send and SNS notification and route it to a desired consumer.
- 0 kudos
- 2925 Views
- 1 replies
- 0 kudos
Azure Databricks Repos and HIPAA
Are Repos HIPAA compliant, or is there a plan and timeline to support this? Customer is getting a warning when trying to enable the Repos feature in a HIPAA deployment on Azure Databricks.
- 2925 Views
- 1 replies
- 0 kudos
- 0 kudos
There is a plan to support this. For timeline, please reach out to your Databricks account team.
- 0 kudos
- 2153 Views
- 1 replies
- 0 kudos
- 2153 Views
- 1 replies
- 0 kudos
- 0 kudos
Unfortunately this is not possible. The default user workspace name will be the user's email address.
- 0 kudos
- 1601 Views
- 1 replies
- 0 kudos
What do I need to think about for Disaster Recovery planning?
I am working on a disaster recovery plan for my environment which includes Databricks. Where do I start with my planning? What all do I need to consider when building a DR plan?
- 1601 Views
- 1 replies
- 0 kudos
- 0 kudos
Depending on your RPO/RTOs there are different recovery solution strategies that could be considered (active/passive, active/active) for Databricks deployments. A detailed explanation of these approaches are mentioned here
- 0 kudos
- 1698 Views
- 1 replies
- 0 kudos
Can you use credential passthrough for users running jobs?
I would like it if I could make it so that the credentials of the user who initiates a job are used as the credentials for the job run. Is this possible?
- 1698 Views
- 1 replies
- 0 kudos
- 0 kudos
Is this in Azure? If so, it is not supported currently. https://docs.microsoft.com/en-us/azure/databricks/security/credential-passthrough/adls-passthrough
- 0 kudos
Join Us as a Local Community Builder!
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now-
Access control
1 -
Apache spark
1 -
Azure
7 -
Azure databricks
5 -
Billing
2 -
Cluster
1 -
Compliance
1 -
Data Ingestion & connectivity
5 -
Databricks Runtime
1 -
Databricks SQL
2 -
DBFS
1 -
Dbt
1 -
Delta Sharing
1 -
DLT Pipeline
1 -
GA
1 -
Gdpr
1 -
Github
1 -
Partner
47 -
Public Preview
1 -
Service Principals
1 -
Unity Catalog
1 -
Workspace
2
- « Previous
- Next »
| User | Count |
|---|---|
| 108 | |
| 37 | |
| 34 | |
| 25 | |
| 24 |