cancel
Showing results for 
Search instead for 
Did you mean: 
Community Platform Discussions
Connect with fellow community members to discuss general topics related to the Databricks platform, industry trends, and best practices. Share experiences, ask questions, and foster collaboration within the community.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Rishabh-Pandey
by Esteemed Contributor
  • 1485 Views
  • 5 replies
  • 0 kudos

Resolved! Issues with Content Writing on Databricks Community

Hi @Sujitha , @Rishabh_Tiwari ,I wanted to bring to your attention that whenever I’m writing content on Databricks, I often encounter errors due to invalid HTML. Additionally, some terms seem to be prohibited by the Databricks community, which is puz...

  • 1485 Views
  • 5 replies
  • 0 kudos
Latest Reply
Rishabh_Tiwari
Databricks Employee
  • 0 kudos

@Rishabh-Pandey I understand. Please be assured that I am actively working on this and tracking these posts to update our filter on a regular basis. If you come across something similar again, feel free to tag me, and I'll take care of that. 

  • 0 kudos
4 More Replies
Rishabh-Pandey
by Esteemed Contributor
  • 537 Views
  • 0 replies
  • 1 kudos

Live, Virtual Workshop How to build a Golden Data Warehouse in Financial Services with Databricks

Reasons to join: Most Financial Services organizations have major on-prem investments. You can use that as your starting point to activate your organization on gold-level insights in the cloud.Providing a path to easier and quicker migration to the c...

  • 537 Views
  • 0 replies
  • 1 kudos
kro
by New Contributor
  • 491 Views
  • 0 replies
  • 0 kudos

OCRmyPDF in Databricks

Hello,Do any of you have experience with using OCRmyPDF in Databricks? I have tried to install it in various was with different versions, but my notebook keep crashing with the error:The Python process exited with exit code 139 (SIGSEGV: Segmentation...

Community Platform Discussions
ocr
ocrmypdf
pdf
segmentation fault
tesseract
  • 491 Views
  • 0 replies
  • 0 kudos
Sourav789027
by New Contributor II
  • 362 Views
  • 1 replies
  • 1 kudos

Databricks Certifications

Hello Everyone , My name is Sourav Das. I am from Kolkata, currently working as Azure Data Engineer in Cognizant.I have cleared multiple databricks certifications(Databricks data engineer associate, databricks data engineer professional, databricks d...

  • 362 Views
  • 1 replies
  • 1 kudos
Latest Reply
gchandra
Databricks Employee
  • 1 kudos

Good luck. You can continue to improve your skills by helping other community members on this platform.

  • 1 kudos
DineshReddyN
by New Contributor II
  • 1096 Views
  • 5 replies
  • 0 kudos

Filestore endpoint not visible in Databricks community edition

In community edition of Databricks after multiple attempts of enable, refreshes, unable to navigate to File store endpoint.Under catalog it is not visible 

DineshReddyN_0-1729153172344.png
Community Platform Discussions
Databricks Community Edition
filestore
GUI
  • 1096 Views
  • 5 replies
  • 0 kudos
Latest Reply
gchandra
Databricks Employee
  • 0 kudos

Follow these alternate solutions. https://community.databricks.com/t5/data-engineering/databricks-community-edition-dbfs-alternative-solutions/td-p/94933

  • 0 kudos
4 More Replies
Sudheer2
by New Contributor III
  • 322 Views
  • 1 replies
  • 0 kudos

Migrating ML Model Experiments Using Python REST APIs

Hi everyone,I’m looking to migrate ML model experiments from a source Databricks workspace to a target workspace. Specifically, I want to use Python and the available REST APIs for this process.Can anyone help me on this!Thanks in advance!

  • 322 Views
  • 1 replies
  • 0 kudos
Latest Reply
gchandra
Databricks Employee
  • 0 kudos

You can use https://github.com/mlflow/mlflow-export-import  utility. The example given below doesn't use Python but uses CLI and CICD pipeline to do the same.  https://medium.com/@gchandra/databricks-copy-ml-models-across-unity-catalog-metastores-188...

  • 0 kudos
abubakar-saddiq
by New Contributor
  • 2008 Views
  • 2 replies
  • 1 kudos

How to Pass Dynamic Parameters (e.g., Current Date) in Databricks Workflow UI?

I'm setting up a job in the Databricks Workflow UI and I want to pass a dynamic parameter, like the current date (run_date), each time the job runs.In Azure Data Factory, I can use expressions like @utcnow() to calculate this at runtime. However, I w...

  • 2008 Views
  • 2 replies
  • 1 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 1 kudos

As szymon mentioned, dynamic parameter values exist, but the functionality is still far from what Data Factory has to offer.I am pretty sure though that this will be extended.So for the moment I suggest you do the value derivation in data factory, an...

  • 1 kudos
1 More Replies
oleprince
by New Contributor
  • 5248 Views
  • 1 replies
  • 0 kudos

Delta table definition - Identity column

Hello,Would anyone know if it is possible to create a delta table using Python that includes a column that is generated by default as identity (identity column for which the value inserted can be manually overriden)?There seems to be a way to create ...

  • 5248 Views
  • 1 replies
  • 0 kudos
Latest Reply
gmiguel
Contributor
  • 0 kudos

Hi @oleprince ,As far as I know, it's not possible yet to create tables with Identity columns using pyspark (DeltaTable api). You can create generated columns, but Identity columns are not allowed.The only way to achieve this is through Spark Sql.  

  • 0 kudos
fiverrpromotion
by New Contributor
  • 528 Views
  • 1 replies
  • 0 kudos

Addressing Memory Constraints in Scaling XGBoost and LGBM: A Comprehensive Approach for High-Volume

Scaling XGBoost and LightGBM models to handle exceptionally large datasets—those comprising billions to tens of billions of rows—presents a formidable computational challenge, particularly when constrained by the limitations of in-memory processing o...

  • 528 Views
  • 1 replies
  • 0 kudos
Latest Reply
earntodiessaz
New Contributor II
  • 0 kudos

Well, that's a superb article! Thank you for this great information, you write very well which I like very much. I am really impressed by your post. run 3

  • 0 kudos
Chris_Shehu
by Valued Contributor III
  • 1282 Views
  • 1 replies
  • 1 kudos

Feature Request: GUI: Additional Collapse options

When you're using a very large notebook sometimes it gets frustrating scrolling through all the code blocks. It would be nice to have a few additional options to make this easier. 1) Add a collapse all code cells button to the top.2) Add a collapse a...

Community Platform Discussions
Enhancement
Feature
GUI
Request
  • 1282 Views
  • 1 replies
  • 1 kudos
Latest Reply
fdawoud
New Contributor II
  • 1 kudos

this feature please

  • 1 kudos
qwerty3
by Contributor
  • 672 Views
  • 1 replies
  • 0 kudos

Resolved! Does a queued databricks job incur cost?

Does a queued databricks job incur cost?

  • 672 Views
  • 1 replies
  • 0 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 0 kudos

Hi @qwerty3 ,No, it does not. When a job is queued (waiting for an available cluster or resources), there is no compute usage, so no charges are incurred for Databricks units (DBUs) or cloud infrastructure (VMs). The queued state is essentially a wai...

  • 0 kudos
tejaswi24
by New Contributor III
  • 2001 Views
  • 11 replies
  • 1 kudos

Resolved! databricks Asset Bundle

i have come accross a documentation on asset bundles long back whcih states that when you typedatabricks bundle initit gives us option to choose a project type. But i see the below error when i do that i see the below erroris there a way, i can take ...

tejaswi24_0-1727696851777.png
Community Platform Discussions
Databricks Asset Bundle
databricks bundle
  • 2001 Views
  • 11 replies
  • 1 kudos
Latest Reply
gchandra
Databricks Employee
  • 1 kudos

Bash

  • 1 kudos
10 More Replies
DW
by New Contributor II
  • 1113 Views
  • 1 replies
  • 2 kudos

column mask on <tinyint>Y columns gives error

My table breaks when I try to mask a column with a name like `<tinyint>Y` -- Create a table with a masked column> CREATE FUNCTION mask_int_col(col_val INTEGER) RETURN CASE WHEN is_member('HumanResourceDept') THEN col_val ELSE CAST(NULL as INTEGER) EN...

  • 1113 Views
  • 1 replies
  • 2 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 2 kudos

Hi @DW ,I have replicated your scenario and encountered the same error when applying a column mask to a column named 1Y in Databricks SQL.In short, it makes sense simply to follow Databricks documentation and use the SQL naming conventions, so that c...

  • 2 kudos
abueno
by New Contributor III
  • 492 Views
  • 1 replies
  • 0 kudos

Databricks Pyspark filter several columns with similar criteria

I am querying a table from the Databricks Catalog which I have to filter several columns with the same criteria.  below is what I have created so far.  I have 10 columns that I have filter with a set of criteria from (dx_list1) and another 10 that I ...

  • 492 Views
  • 1 replies
  • 0 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 0 kudos

Hi @abueno ,As I understand the logic you want to implement is:1. For every pair of columns:First Column (DX_i): Must be in dx_list1Second Column (DX_{i+1}): Must be in dx_list22. The condition for each pair is:col('DX_i').isin(dx_list1) OR col('DX_{...

  • 0 kudos
ayush19
by New Contributor III
  • 1643 Views
  • 3 replies
  • 1 kudos

How to retrieve Spark Session inside java jar library installed on Cluster

I have a java app in form of jar package. This jar is installed on a Databricks cluster. This jar package reads and writes to few tables in databricks. In order to achieve that, I need SparkSession available in the code. Given that spark session is a...

  • 1643 Views
  • 3 replies
  • 1 kudos
Latest Reply
IslaGray
New Contributor II
  • 1 kudos

Thanks for the update, I will try it too.

  • 1 kudos
2 More Replies