- 1376 Views
- 1 replies
- 1 kudos
Resolved! Addressing Memory Constraints in Scaling XGBoost and LGBM: A Comprehensive Approach for High-Volume
Scaling XGBoost and LightGBM models to handle exceptionally large datasets—those comprising billions to tens of billions of rows—presents a formidable computational challenge, particularly when constrained by the limitations of in-memory processing o...
- 1376 Views
- 1 replies
- 1 kudos
- 1 kudos
Hi @fiverrpromotion, As you mention, scaling XGBoost and LightGBM for massive datasets has its challenges, especially when trying to preserve critical training capabilities such as early stopping and handling of sparse features / high-cardinality cat...
- 1 kudos
- 1318 Views
- 1 replies
- 0 kudos
Resolved! Problem with ray train and Databricks Notebook (Strange dbutils error)
Hi everyone,I'm running some code to train a multimodal Hugging Face model with SFTTrainer and TorchTrainer to use all GPU workers. When trying to execute trainer.fit() it gives me a dbutils serialization error,even I am not using dbutils directly in...
- 1318 Views
- 1 replies
- 0 kudos
- 0 kudos
JavierS - The dbutils serialization error occurs in your code because dbutils is only available on the Databricks driver node and cannot be pickled or transferred to Spark or Ray worker nodes. This error can appear even if your code doesn't directly...
- 0 kudos
- 760 Views
- 2 replies
- 1 kudos
Resolved! Using merge Schema with spark.read.csv for inconsistent schemas
The Problem:A common data engineering challenge is reading a directory of CSV files where the schemas are inconsistent. For example, some files might have columns in a different order, or be missing certain columns altogether. The standard behavior o...
- 760 Views
- 2 replies
- 1 kudos
- 1 kudos
Hey @JaydeepKhatri here are some helpful points to consider: Is this an officially supported, enhanced feature of the Databricks CSV reader? Based on internal research, this appears to be an undocumented “feature” of Spark running on Databricks. ...
- 1 kudos
- 4240 Views
- 10 replies
- 0 kudos
Who's hiring? Latest Job Postings from the Databricks Community!
More than 10,000 organizations worldwide — including Block, Comcast, Condé Nast, Rivian, Shell and over 60% of the Fortune 500 — rely on the Databricks Data Intelligence Platform to take control of their data and put it to work with AI. Use this thre...
- 4240 Views
- 10 replies
- 0 kudos
- 0 kudos
Job Title - Engineering Data Analyst LeadCompany name- Wartsila Location - Mumbai Required years of experience- 8 years Link to apply - https://careers.wartsila.com/job-invite/46999/Job description;ResponsibilitiesTransform raw data into meaningful i...
- 0 kudos
- 3422 Views
- 5 replies
- 0 kudos
Talend ETL code to Databricks
Hi Team,What is the best way to transfer Talend ETL code to Databricks and what are the best methods/practices for migrating Talend ETL's to Databricks (notebook, code conversion/migration strategy, workflow's etc)?Regards,Janga
- 3422 Views
- 5 replies
- 0 kudos
- 0 kudos
Hi @Phani1 ,Were you able to complete the migration from Talend to Databricks? If yes, please do let me know the process / tools that you have used for the migration and the challenges that you have encountered. ThanksVamsi.
- 0 kudos
- 1006 Views
- 6 replies
- 5 kudos
Resolved! Databricks metastore
I have configured a new databricks workspace through my personal azure account. I cannot see the metastore. Since this is my personal account, I am the admin. Can you please guide me on how to enable the metastore. Thank you.
- 1006 Views
- 6 replies
- 5 kudos
- 5 kudos
Hello @anipar Good day!!Can you please confirm again if you can see the metastore here?Log in to your Databricks workspace.Click your username in the top bar and select "Manage Account" to open the account console AZURE (https://accounts.azuredatabri...
- 5 kudos
- 291 Views
- 1 replies
- 1 kudos
Resolved! Dashboard embed: dashboard id is missing in token claim
We're trying to embed a dashboard following the example at https://learn.microsoft.com/en-us/azure/databricks/dashboards/embedding/external-embed#step-4-use-the-example-app-to-authenticate-and-generate-tokens The embed is displayed but with an error ...
- 291 Views
- 1 replies
- 1 kudos
- 1 kudos
Nevermind. I accidentally removed the external_viewer_id and external_value parameters. After adding them back it works as expected.
- 1 kudos
- 329 Views
- 0 replies
- 3 kudos
Understanding Databricks Clean Rooms: Design, Access, and Cost Model
Hi everyone,I wanted to share some insights and clarifications around Databricks Clean Rooms, especially regarding how they are designed and how costs are calculated — something that can be a bit confusing when onboarding collaborators.What Are Clean...
- 329 Views
- 0 replies
- 3 kudos
- 247 Views
- 1 replies
- 0 kudos
Resolved! Signed up for Customer Academy instead of Partner
Hello. Yesterday, I signed up for an account under my company email as a customer account. However, my intent was to sign up for a partner account and now it won't let me switch back. I sent a ticket to the academy team but have not received a reply ...
- 247 Views
- 1 replies
- 0 kudos
- 0 kudos
Hello @ericmedina! For issues with Academy access, the correct step is to raise a ticket with the Databricks support team. Since you’ve already submitted one, please allow some time for them to respond.
- 0 kudos
- 2947 Views
- 6 replies
- 1 kudos
Resolved! Spark connect client and server versions should be same for executing UDFs
I am trying to execute pandas UDF in databricks. It gives me the following error on serverless compute,File /local_disk0/.ephemeral_nfs/envs/pythonEnv-b11ff17c-9b25-4ccb-927d-06a7d1ca7221/lib/python3.11/site-packages/pyspark/sql/connect/client/core.p...
- 2947 Views
- 6 replies
- 1 kudos
- 1 kudos
@chinmay0924 - You can change the serverless client image by selecting the environment panel in a Databricks Notebook (on the right pane) or in the "Environment and Libraries" section while configuring a Databricks Job Task. The set of available serv...
- 1 kudos
- 523 Views
- 1 replies
- 0 kudos
Exam Suspended After Restarting the Exam — Ticket ID: 00744386
Hi Team,I am writing to raise a concern regarding my Databricks Certified Data Engineer Professional exam (Ticket ID: 00744386).During my exam session, the support representative initially requested me to show the entire room, including my bed, whi...
- 523 Views
- 1 replies
- 0 kudos
- 0 kudos
Hello @rajkishore028,Thank you for filing a ticket with our support team, Support team will respond shortly. Please note that we cannot provide support or handle exam suspensions via community. Thanks & Regards,@cert-ops
- 0 kudos
- 332 Views
- 1 replies
- 2 kudos
Resolved! How to update comments and constraints on Streaming Tables created by DLT outside the pipeline?
I have a Delta Live Tables (DLT) pipeline that creates a streaming table (e.g., dbrtdflite_catalog.demotest_silver.customer_profile).I want to update table-level and column-level comments as well as constraints on this table.When I try commands like:...
- 332 Views
- 1 replies
- 2 kudos
- 2 kudos
Hi @Nexusss7 You can add a comment in the delta live tables, either the MV or the streaming table, in the tag - @Dlt.table() @Dlt.table( comment = "Delta live tables comment" )Here is the syntax for SQL: https://docs.databricks.com/aws/en/d...
- 2 kudos
- 768 Views
- 1 replies
- 1 kudos
Blackduck scanning on Databricks Workflow
Does anyone know if its compatible scan in blackduck your json based files from Workflows? At least, when its come to the notebook its compatible as blackduck detects python based files but i am wondering why can workflow be scanned as well.
- 768 Views
- 1 replies
- 1 kudos
- 1 kudos
Hi @fjrodriguez , Have you found your answer already? I hope I understand your question correctly. I think Databricks Asset Bundles (or Terraform) can be a useful tool to be added here, if you haven't looked at it already. Yes, you can and should s...
- 1 kudos
- 7105 Views
- 5 replies
- 1 kudos
Databricks Job Failure + Service now Integration
Hi Team,Could you please suggest how to raise the service now ticket, in case of Databricks job failure?Regards ,Phanindra
- 7105 Views
- 5 replies
- 1 kudos
- 1 kudos
JSON response from Databricks to ServiceNow, doesn't give enough details. Is there a way we can send customized json with additional details.
- 1 kudos
- 3711 Views
- 8 replies
- 0 kudos
Unable to install the Databricks in AWS
I have followed the instructions given in the following link https://docs.databricks.com/en/admin/workspace/quick-start.html#use-aws-quickstart-to-create-a-databricks-workspace%22There was an error in CFT while deploying the workspace. Please help[ER...
- 3711 Views
- 8 replies
- 0 kudos
- 0 kudos
Hi @rseshadri Please check the following two points:STS Endpoint Activation: Confirm whether the STS endpoint for the us-west-2 region is active. You can verify this from the IAM account settings.Resource Limits: Ensure that the limits for resources ...
- 0 kudos
Join Us as a Local Community Builder!
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now-
.CSV
1 -
Access Data
2 -
Access Databricks
1 -
Access Delta Tables
2 -
Account reset
1 -
ADF Pipeline
1 -
ADLS Gen2 With ABFSS
1 -
Advanced Data Engineering
1 -
AI
2 -
Analytics
1 -
Apache spark
1 -
Apache Spark 3.0
1 -
Api Calls
1 -
API Documentation
3 -
Architecture
1 -
asset bundle
1 -
Asset Bundles
3 -
Auto-loader
1 -
Autoloader
4 -
AWS security token
1 -
AWSDatabricksCluster
1 -
Azure
6 -
Azure data disk
1 -
Azure databricks
14 -
Azure Databricks SQL
6 -
Azure databricks workspace
1 -
Azure Unity Catalog
5 -
Azure-databricks
1 -
AzureDatabricks
1 -
AzureDevopsRepo
1 -
Big Data Solutions
1 -
Billing
1 -
Billing and Cost Management
1 -
Blackduck
1 -
Bronze Layer
1 -
Certification
3 -
Certification Exam
1 -
Certification Voucher
3 -
CICDForDatabricksWorkflows
1 -
Cloud_files_state
1 -
CloudFiles
1 -
Cluster
3 -
Cluster Init Script
1 -
Community Edition
3 -
Community Event
1 -
Community Group
2 -
Community Members
1 -
Compute
3 -
Compute Instances
1 -
conditional tasks
1 -
Connection
1 -
Contest
1 -
Credentials
1 -
Custom Python
1 -
CustomLibrary
1 -
Data
1 -
Data + AI Summit
1 -
Data Engineering
3 -
Data Explorer
1 -
Data Ingestion & connectivity
1 -
Data Processing
1 -
Databrick add-on for Splunk
1 -
databricks
2 -
Databricks Academy
1 -
Databricks AI + Data Summit
1 -
Databricks Alerts
1 -
Databricks Assistant
1 -
Databricks Certification
1 -
Databricks Cluster
2 -
Databricks Clusters
1 -
Databricks Community
10 -
Databricks community edition
3 -
Databricks Community Edition Account
1 -
Databricks Community Rewards Store
3 -
Databricks connect
1 -
Databricks Dashboard
3 -
Databricks delta
2 -
Databricks Delta Table
2 -
Databricks Demo Center
1 -
Databricks Documentation
3 -
Databricks genAI associate
1 -
Databricks JDBC Driver
1 -
Databricks Job
1 -
Databricks Lakehouse Platform
6 -
Databricks Migration
1 -
Databricks Model
1 -
Databricks notebook
2 -
Databricks Notebooks
3 -
Databricks Platform
2 -
Databricks Pyspark
1 -
Databricks Python Notebook
1 -
Databricks Repo
1 -
Databricks Runtime
1 -
Databricks SQL
5 -
Databricks SQL Alerts
1 -
Databricks SQL Warehouse
1 -
Databricks Terraform
1 -
Databricks UI
1 -
Databricks Unity Catalog
4 -
Databricks Workflow
2 -
Databricks Workflows
2 -
Databricks workspace
3 -
Databricks-connect
1 -
databricks_cluster_policy
1 -
DatabricksJobCluster
1 -
DataCleanroom
1 -
DataDays
1 -
Datagrip
1 -
DataMasking
2 -
DataVersioning
1 -
dbdemos
2 -
DBFS
1 -
DBRuntime
1 -
DBSQL
1 -
DDL
1 -
Dear Community
1 -
deduplication
1 -
Delt Lake
1 -
Delta Live Pipeline
3 -
Delta Live Table
5 -
Delta Live Table Pipeline
5 -
Delta Live Table Pipelines
4 -
Delta Live Tables
7 -
Delta Sharing
2 -
deltaSharing
1 -
Deny assignment
1 -
Development
1 -
Devops
1 -
DLT
10 -
DLT Pipeline
7 -
DLT Pipelines
5 -
Dolly
1 -
Download files
1 -
Dynamic Variables
1 -
Engineering With Databricks
1 -
env
1 -
ETL Pipelines
1 -
External Sources
1 -
External Storage
2 -
FAQ for Databricks Learning Festival
2 -
Feature Store
2 -
Filenotfoundexception
1 -
Free trial
1 -
GCP Databricks
1 -
GenAI
1 -
Getting started
2 -
Google Bigquery
1 -
HIPAA
1 -
Hubert Dudek
2 -
import
1 -
Integration
1 -
JDBC Connections
1 -
JDBC Connector
1 -
Job Task
1 -
Lineage
1 -
LLM
1 -
Login
1 -
Login Account
1 -
Machine Learning
3 -
MachineLearning
1 -
Materialized Tables
2 -
Medallion Architecture
1 -
Migration
1 -
ML Model
2 -
MlFlow
2 -
Model Training
1 -
Module
1 -
Networking
1 -
Notebook
1 -
Onboarding Trainings
1 -
OpenAI
1 -
Pandas udf
1 -
Permissions
1 -
personalcompute
1 -
Pipeline
2 -
Plotly
1 -
PostgresSQL
1 -
Pricing
1 -
Pyspark
1 -
Python
5 -
Python Code
1 -
Python Wheel
1 -
Quickstart
1 -
Read data
1 -
Repos Support
1 -
Reset
1 -
Rewards Store
2 -
Schedule
1 -
Serverless
3 -
serving endpoint
1 -
Session
1 -
Sign Up Issues
2 -
Spark Connect
1 -
sparkui
2 -
Splunk
2 -
SQL
8 -
Summit23
7 -
Support Tickets
1 -
Sydney
2 -
Table Download
1 -
Tags
1 -
terraform
1 -
Training
2 -
Troubleshooting
1 -
Unity Catalog
4 -
Unity Catalog Metastore
2 -
Update
1 -
user groups
1 -
Venicold
3 -
Voucher Not Recieved
1 -
Watermark
1 -
Weekly Documentation Update
1 -
Weekly Release Notes
2 -
Women
1 -
Workflow
2 -
Workspace
3
- « Previous
- Next »
| User | Count |
|---|---|
| 133 | |
| 116 | |
| 56 | |
| 42 | |
| 34 |