cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

ChristianRRL
by Contributor II
  • 492 Views
  • 0 replies
  • 0 kudos

Source to Bronze Organization + Partition

Hi there, I hope I have what is effectively a simple question. I'd like to ask for a bit on guidance if I am structuring my source-to-bronze auto loader data properly. Here's what I have currently:/adls_storage/<data_source_name>/<category>/autoloade...

  • 492 Views
  • 0 replies
  • 0 kudos
pablobd
by Contributor
  • 1927 Views
  • 3 replies
  • 0 kudos

Install python package from private repo [CodeArtifact]

As part of my MLOps stack, I have developed a few packages which are the published to a private AWS CodeArtifact repo. How can I connect the AWS CodeArtifact repo to databricks? I want to be able to add these packages to the requirements.txt of a mod...

  • 1927 Views
  • 3 replies
  • 0 kudos
Latest Reply
pablobd
Contributor
  • 0 kudos

One way to do it is to run this line before installing the dependencies:pip config set site.index-url https://aws:$CODEARTIFACT_AUTH_TOKEN@my_domain-111122223333.d.codeartifact.region.amazonaws.com/pypi/my_repo/simple/But can we add this in MLFlow?

  • 0 kudos
2 More Replies
shkelzeen
by New Contributor II
  • 879 Views
  • 2 replies
  • 0 kudos

Databricks JDBC driver multi query in one request.

Can I run multi query in one command using databricks JDBC driver and would databricks execute one query faster then running multi queries in one script?  

  • 879 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @shkelzeen,  You can run multiple queries in one command using Databricks JDBC driver by using the operator to chain multiple SQL statements together.    This will execute two queries in one command and display the results in separate tables.   ...

  • 0 kudos
1 More Replies
harvey-c
by New Contributor III
  • 544 Views
  • 1 replies
  • 0 kudos

DLT pipeline access external location with abfss protocol was failed

Dear Databricks Community Members:The symptom: The DLT pipeline was failed with the error message:  Failure to initialize configuration for storage account storageaccount.dfs.core.windows.net: Invalid configuration value detected for fs.azure.account...

  • 544 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @harvey-c,    One common issue is using the container names instead of the storage accounts in the URL, or including the container names in the configs. Ensure you use the correct storage account name and key in your code and settings.Another comm...

  • 0 kudos
ChristianRRL
by Contributor II
  • 1039 Views
  • 3 replies
  • 1 kudos

Auto Loader Use Case Question - Centralized Dropzone to Bronze?

Good day,I am trying to use Auto Loader (potentially extending into DLT in the future) to easily pull data coming from an external system (currently located in a single location) and organize it and load it respectively. I am struggling quite a bit a...

  • 1039 Views
  • 3 replies
  • 1 kudos
Latest Reply
ChristianRRL
Contributor II
  • 1 kudos

Quick follow-up on this @Kaniz_Fatma (or to anyone else in the Databricks multi-verse who is able to help clarify this case).I understand that the proposed solution would work for a "one-to-one" case where many files are landing in a specific dbfs pa...

  • 1 kudos
2 More Replies
Cosmin
by New Contributor II
  • 1162 Views
  • 3 replies
  • 0 kudos

Fail to write large dataframe

Hi all, we have a issue while trying to write a quite large data frame, close to 35 million records. We try to write it as parquet and also table and none work. But writing a small chink (10k records) is working. Basically we have some text on which ...

Cosmin_2-1702640369404.png
  • 1162 Views
  • 3 replies
  • 0 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 0 kudos

That could work, but you will have to create a UDF.Check this SO topic for more info

  • 0 kudos
2 More Replies
vrajesh123
by New Contributor II
  • 2690 Views
  • 2 replies
  • 0 kudos

Resolved! Webassessor Secure Browser will not Launch during exam.

Hello - I registered for the Databricks Data Engineering Associate Certification exam.  I hit an issue, their Secure browser would not launch, it just crashed - the only thing I could see in a flash is "bad request" and poof its gone.  Spend over 2 h...

Get Started Discussions
Certification Exam
Issues
Secure Browser
  • 2690 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @vrajesh123,    Thank you for posting your concern on Community! To expedite your request, please list your concerns on our ticketing portal. Our support staff would be able to act faster on the resolution (our standard resolution time is 24-48 ho...

  • 0 kudos
1 More Replies
sudhanshu1
by New Contributor III
  • 7642 Views
  • 3 replies
  • 2 kudos

Python file testing using pytest

Hi All,I have a requirement in my project, where we will be writing some python code inside databricks . Please note we will not be using pyspark . It will plain pythin with polars.I am looking into ho to create test files for main file. Below is sim...

  • 7642 Views
  • 3 replies
  • 2 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 2 kudos

Hi @sudhanshu1 , The error you're encountering is due to the fact that pytest is looking for a fixture named 'name' which is not defined. Fixtures in pytest are functions that are run before each test function to which it is applied. Fixtures are us...

  • 2 kudos
2 More Replies
pablobd
by Contributor
  • 1842 Views
  • 2 replies
  • 1 kudos

Resolved! Asset bundle build and deploy python wheel with versions

Hi all,I was able to deploy a wheel to the /Shared/ folder from a repository in Gitlab with asset bundles. The databricks.yml looks something like this.artifacts:  default:    type: whl    build: poetry build    path: .  targets:    workspace:      h...

  • 1842 Views
  • 2 replies
  • 1 kudos
Latest Reply
pablobd
Contributor
  • 1 kudos

Finally I decided to use AWS Code Artifact and mirror the PyPI, which I think it's a bit cleaner. But your solution looks good too. Thanks!

  • 1 kudos
1 More Replies
Phani1
by Valued Contributor
  • 589 Views
  • 1 replies
  • 0 kudos

informatica jobs from data bricks

Hi TeamHow can we call informatica jobs from data bricks? could you please suggest on this.Regards,Phanindra

Get Started Discussions
informatica jobs
  • 589 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @Phani1, Informatica and Databricks have a partnership that allows for seamless integration between the two platforms.   Here’s a high-level overview of how you can use Informatica with Databricks:   Data Ingestion: You can ingest data directly in...

  • 0 kudos
leelee3000
by New Contributor III
  • 843 Views
  • 2 replies
  • 0 kudos

time travel with DLT

Needed some help with Time Travel with Delta Live tables   We were trying to figure out if we can go in and alter the history on this table, and what would happen to data that we mass upload?  By this we mean we have data from the past that we would ...

  • 843 Views
  • 2 replies
  • 0 kudos
Latest Reply
BigRoux
New Contributor III
  • 0 kudos

Delta Live Tables leverage Delta Lake, or Delta Tables.  Delta tables, through transactions (e.g. insert, update, delete, merges, optimization) create versions of said Delta Table.  Once a version is created it cannot be altered, it is immutable.  Yo...

  • 0 kudos
1 More Replies
Phani1
by Valued Contributor
  • 860 Views
  • 2 replies
  • 0 kudos

Upgrade Spark version 3.2 to 3.4+

Hi Team,We would like to upgrade from Spark version 3.2 to 3.4+ (Databricks Runtime - 10.4 to 12.2/13.3)We would like to understand how complex upgradation is this and challenges which we face? what are the technical steps and precautions we need to ...

  • 860 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @Phani1,   Check Databricks Runtime release notes for any specific changes, improvements, or deprecations that might impact your workloads. Our End-of-Year Community Survey is here! Please take a few moments to complete the survey. Your feedback m...

  • 0 kudos
1 More Replies
RamanP9404
by New Contributor
  • 1232 Views
  • 1 replies
  • 0 kudos

Spark Streaming Issues while performing left join

Hi team,I'm struck in a Spark Structured streaming use-case.Requirement: To read two streaming data frames, perform a left join on it and display the results. Issue: While performing a left join, the resultant data frame contains only rows where ther...

  • 1232 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @RamanP9404, In Spark Structured Streaming, watermarking is essential for handling late data and ensuring correctness in stream-stream joins. When performing a left outer join, both the left and right datasets must be watermarked with a time const...

  • 0 kudos
Phani1
by Valued Contributor
  • 1562 Views
  • 1 replies
  • 0 kudos

Customer Managed Keys in Databricks (AWS)

Hi Databricks Team,Could you please provide me the detailed steps on how to be enabled customer managed keys in databricks (AWS) Account, if there is any video on it that would be great helpful.Regards,Phanindra

  • 1562 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @Phani1, Enabling customer-managed keys (CMKs) in Databricks on AWS involves securing and controlling access to encrypted data.    Here are the steps to set up CMKs:   Prerequisites: Ensure your Databricks workspace is on the E2 version of the Dat...

  • 0 kudos
Join 100K+ Data Experts: Register Now & Grow with Us!

Excited to expand your horizons with us? Click here to Register and begin your journey to success!

Already a member? Login and join your local regional user group! If there isn’t one near you, fill out this form and we’ll create one for you to join!

Labels
Top Kudoed Authors