- 11523 Views
- 6 replies
- 9 kudos
Notebook scrolling as I select
When I select text in a notebook cell the whole notebook scrolls up as a select. This happens when I use the mouse wheel and with shift+arrow key. It varies by cell - happens in some cells, but not other cells within the same notebook. When I refresh...
- 11523 Views
- 6 replies
- 9 kudos
- 9 kudos
I'm relieved to know that I'm not the only one experiencing this issue.Please, address this as soon as possible. It's significantly impacting my productivity.
- 9 kudos
- 1084 Views
- 1 replies
- 3 kudos
Resolved! How do i fix QB Error code 6000 301?
When I try to access my company files, I keep getting the error 6000 301 in QB. Please assist me in fixing this mistake.
- 1084 Views
- 1 replies
- 3 kudos
- 3 kudos
Are you facing QB error 6000 301 while trying to log in to your company file in QB Desktop and don’t know what should be done next? If yes, then you should not panic at all because through this post I am going to tell you everything you need to know ...
- 3 kudos
- 1508 Views
- 2 replies
- 0 kudos
Resolved! Create Storage Credential 500 Response
I'm trying to create storage credentials for an Azure Databricks Connector at the workspace level with a service principal that has the CREATE_STORAGE_CREDENTIAL but is NOT an account admin. For this test, the SP has the owner role on the connector.I...
- 1508 Views
- 2 replies
- 0 kudos
- 0 kudos
Hi @ledbutter , hope you are doing well today! I have gone through the details and this issue might be related to https://github.com/databricks/cli/issues/1080 Please refer to this for more details: https://github.com/databricks/cli/issues/1108 Plea...
- 0 kudos
- 874 Views
- 2 replies
- 0 kudos
Cells' outputs getting appended at each run - Databricks Notebook
Hello Community,I have the following issue. When I am running cells from a notebook, I have the print outputs from the previous cells that are appended to the current print output (meaning running cell 1 gives output 1, running cell 2 gives output 1 ...
- 874 Views
- 2 replies
- 0 kudos
- 0 kudos
This seems to be linked by installing pycaret
- 0 kudos
- 2770 Views
- 2 replies
- 2 kudos
Resolved! Rss feeds for databricks releases
Hi,are there any rss feeds for the databricks platform, sql & runtime releases? We have a big tech stack so it is sometimes hard to keep up with the ever changing technologies. We are using rss feeds to keep up with all of that.Cant find anything for...
- 2770 Views
- 2 replies
- 2 kudos
- 2 kudos
Hi @florianb , You can keep these documents handy for the Platform releases and SQL Runtime releases: Platform Releases: https://docs.databricks.com/en/release-notes/product/index.html SQL Releases: https://docs.databricks.com/en/sql/release-notes/in...
- 2 kudos
- 1116 Views
- 1 replies
- 0 kudos
Getting internal server error while creating a new query definition
Hi, I am trying to create a query definition using API '/api/2.0/preview/sql/queries' in postman but getting internal server error. Below is the snap for the snap. Let me know If I am doing anything wrong here.
- 1116 Views
- 1 replies
- 0 kudos
- 0 kudos
I'm interested to know how this error was resolved. I'm getting an "Internal Server Error" returned when trying to create queries with version 1.36.1 of the Databricks Terraform Provider. The error provides no other information.
- 0 kudos
- 6718 Views
- 2 replies
- 0 kudos
Parallel jobs with individual contexts
I was wondering if someone could help us with implementation here. Our current program will spin up 5 jobs through the Databricks API using the same Databricks cluster but each one needs their own spark context (specifically each one will connect to ...
- 6718 Views
- 2 replies
- 0 kudos
- 0 kudos
you can set up buckets with different credentials, endpoints, and so on.https://docs.databricks.com/en/connect/storage/amazon-s3.html#per-bucket-configuration
- 0 kudos
- 7402 Views
- 0 replies
- 0 kudos
Data Engineer – Databricks - Remote
Data Engineer – Databricks - RemoteApply Here: Job Application for Data Engineer – Databricks at Jenzabar (greenhouse.io)Jenzabar Website: Higher Education Software Solutions - JenzabarFor over four decades, the higher education experts at Jenzabar h...
- 7402 Views
- 0 replies
- 0 kudos
- 10088 Views
- 0 replies
- 1 kudos
Calling all innovators and visionaries! The 2024 Data Team Awards are open for nominations
Each year, we celebrate the amazing customers that rely on Databricks to innovate and transform their organizations — and the world — with the power of data and AI. The nomination form is now open to submit nominations. Nominations will close on Marc...
- 10088 Views
- 0 replies
- 1 kudos
- 2367 Views
- 4 replies
- 0 kudos
Databricks XML - Bypassing rootTag and rowTag
I see the current conversion of dataframe to xml need to be improved.My dataframe schema is a perfect nested schema based on structs but when I create a xml I have the follow issues:1) I can't add elements to root2) rootTag and rowTag are requiredIn ...
- 2367 Views
- 4 replies
- 0 kudos
- 0 kudos
Here is one of the ways to use the struct field name as rowTag: import org.apache.spark.sql.types._ val schema = new StructType().add("Record", new StructType().add("age", IntegerType).add("name", StringType)) val data = Seq(Row(Row(18, "John ...
- 0 kudos
- 1764 Views
- 3 replies
- 0 kudos
The risks of code execution by default on widget change
Taking from my experience, the default action of widgets triggering code execution upon value change poses risks that outweigh the convenience in certain scenarios. While this feature may seem advantageous in some cases, it can lead to unintended con...
- 1764 Views
- 3 replies
- 0 kudos
- 0 kudos
I definitely have to agree with the original point- if you have a notebook that you import, and you touch any widget value you're running code, most likely accidentally. I'd love to see a workspace or user type option where you can change the default...
- 0 kudos
- 1209 Views
- 2 replies
- 1 kudos
databricks spark XML Writer
Hi.I'm trying to generate XML as output base on my nested dataframe. Everything is ok except by I don't know how to add elements to rootTag.I can add elements from rowtag but not in rootTag. Same problems to add attributes to root <books version = "...
- 1209 Views
- 2 replies
- 1 kudos
- 1 kudos
Hi @RobsonNLPT ,Thanks for bringing up your concerns, always happy to help Can you please refer to the below document to read and write the XML files? https://docs.databricks.com/en/query/formats/xml.html Please let me know if this helps and leave a...
- 1 kudos
- 1859 Views
- 1 replies
- 0 kudos
FileAlreadyExistsException error while analyzing table in Notebook
Databricks experts, I'm new to Databricks, and encounter an issue with the ANALYZE TABLE command in the Notebook. I created two tables nyc_taxi and nyc_taxi2, from one csv file.When executing the following command in Notebook, analyze table nyc_taxi2...
- 1859 Views
- 1 replies
- 0 kudos
- 0 kudos
- 0 kudos
- 1486 Views
- 3 replies
- 0 kudos
"PutWithBucketOwnerFullControl" privilege missing for storage configuration
Hi. I've been unable to create workspaces manually for a while now. The error I get is "MALFORMED_REQUEST: Failed storage configuration validation checks: List,Put,PutWithBucketOwnerFullControl,Delete". The storage configuration is on a bucket that ...
- 1486 Views
- 3 replies
- 0 kudos
- 0 kudos
> Yes, it does look like the bucket permissions are not properly set up, but ...To avoid potential misunderstanding: I mean yes the error message does make it sound like the bucket permissions are wrong. I don't meant I found a problem with the ones ...
- 0 kudos
- 8311 Views
- 3 replies
- 1 kudos
stop autoloader with continuous trigger programatically
Hi,I am running autoloader with continuous trigger. How can I stop this trigger during some specific time, only if no data pending and current batch process is complete. How to check how many records pending in queue and current state.Regards,Sanjay
- 8311 Views
- 3 replies
- 1 kudos
- 1 kudos
Hello, I am new here, Can I ask a question?
- 1 kudos
Connect with Databricks Users in Your Area
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group-
AI Summit
4 -
Azure
2 -
Azure databricks
2 -
Bi
1 -
Certification
1 -
Certification Voucher
2 -
Community
7 -
Community Edition
3 -
Community Members
1 -
Community Social
1 -
Contest
1 -
Data + AI Summit
1 -
Data Engineering
1 -
Databricks Certification
1 -
Databricks Cluster
1 -
Databricks Community
8 -
Databricks community edition
3 -
Databricks Community Rewards Store
3 -
Databricks Lakehouse Platform
5 -
Databricks notebook
1 -
Databricks Office Hours
1 -
Databricks Runtime
1 -
Databricks SQL
4 -
Databricks-connect
1 -
DBFS
1 -
Dear Community
1 -
Delta
9 -
Delta Live Tables
1 -
Documentation
1 -
Exam
1 -
Featured Member Interview
1 -
HIPAA
1 -
Integration
1 -
LLM
1 -
Machine Learning
1 -
Notebook
1 -
Onboarding Trainings
1 -
Python
2 -
Rest API
10 -
Rewards Store
2 -
Serverless
1 -
Social Group
1 -
Spark
1 -
SQL
8 -
Summit22
1 -
Summit23
5 -
Training
1 -
Unity Catalog
3 -
Version
1 -
VOUCHER
1 -
WAVICLE
1 -
Weekly Release Notes
2 -
weeklyreleasenotesrecap
2 -
Workspace
1
- « Previous
- Next »