cancel
Showing results for 
Search instead for 
Did you mean: 
Machine Learning
Dive into the world of machine learning on the Databricks platform. Explore discussions on algorithms, model training, deployment, and more. Connect with ML enthusiasts and experts.
cancel
Showing results for 
Search instead for 
Did you mean: 

Exploring Serverless Features in Databricks for ML Use Cases

antonionuzzo
New Contributor II

Hello, 

I need to develop some ML use case. I would like to understand if the serverless functionality unlocks any additional features or if it is mandatory for certain capabilities.

Thank you!

1 ACCEPTED SOLUTION

Accepted Solutions

BigRoux
Databricks Employee
Databricks Employee
Serverless functionality in Databricks is not mandatory for utilizing machine learning (ML) capabilities. However, it does unlock specific benefits and features that can enhance certain workflows. Here’s how serverless compute can add value, based on the context:
  1. Performance and Scalability:
    • Serverless compute allows for fast startup times and automatic scalability, which is particularly useful for ML workloads involving exploratory experiments or interactive use cases where efficiency is key.
  2. Cost Optimization:
    • Serverless compute operates in a cost-optimized mode for workflows, notebooks, and Delta Live Tables, reducing costs when resources are not actively in use. This can particularly benefit intermittent ML workloads.
  3. Enhanced Security and Governance:
    • Serverless environments include enhanced security features, such as shared security access modes and Unity Catalog integration, which support secure and compliant ML workflows.
  4. Separating Responsibility:
    • Serverless eliminates the need for manually provisioning and managing clusters, allowing data scientists and ML practitioners to focus entirely on their work without requiring support from infrastructure teams.
  5. Developing and Managing ML Models:
    • While serverless compute supports ML model development and deployment, limitations exist for workloads requiring GPUs, certain ML runtime features, or custom data sources. However, Databricks MLtools like MLflow can still be leveraged effectively within serverless environments for experiment tracking and deployment.
  6. Limitations:
    • Specific functionality like Spark UI debugging, certain Spark configurations, and support for GPUs or cluster-scoped libraries (e.g., .jar files) is limited in serverless environments. Ensure these constraints align with your ML use case.
Serverless compute is beneficial but not mandatory for most Databricks ML workflows.

View solution in original post

1 REPLY 1

BigRoux
Databricks Employee
Databricks Employee
Serverless functionality in Databricks is not mandatory for utilizing machine learning (ML) capabilities. However, it does unlock specific benefits and features that can enhance certain workflows. Here’s how serverless compute can add value, based on the context:
  1. Performance and Scalability:
    • Serverless compute allows for fast startup times and automatic scalability, which is particularly useful for ML workloads involving exploratory experiments or interactive use cases where efficiency is key.
  2. Cost Optimization:
    • Serverless compute operates in a cost-optimized mode for workflows, notebooks, and Delta Live Tables, reducing costs when resources are not actively in use. This can particularly benefit intermittent ML workloads.
  3. Enhanced Security and Governance:
    • Serverless environments include enhanced security features, such as shared security access modes and Unity Catalog integration, which support secure and compliant ML workflows.
  4. Separating Responsibility:
    • Serverless eliminates the need for manually provisioning and managing clusters, allowing data scientists and ML practitioners to focus entirely on their work without requiring support from infrastructure teams.
  5. Developing and Managing ML Models:
    • While serverless compute supports ML model development and deployment, limitations exist for workloads requiring GPUs, certain ML runtime features, or custom data sources. However, Databricks MLtools like MLflow can still be leveraged effectively within serverless environments for experiment tracking and deployment.
  6. Limitations:
    • Specific functionality like Spark UI debugging, certain Spark configurations, and support for GPUs or cluster-scoped libraries (e.g., .jar files) is limited in serverless environments. Ensure these constraints align with your ML use case.
Serverless compute is beneficial but not mandatory for most Databricks ML workflows.

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now