- 5996 Views
- 0 replies
- 0 kudos
Hi,Is there any way we can connect glue catalog as well as to hive metastore in the same warehouse?I can create a single instance profile and provide all the required access for buckets or for glue catalog.I tried with the below configuration,spark.s...
The best trial results seem to change every time the same study is rerun. On Microsoft Azure this can be fixed by setting the sampler seed. However this solution doesn't seem to work on Databricks. Does anyone know why that is the case and how to mak...
I know this has probably been asked many times, although I have struggled to find the answer using sql CREATE STREAMING LIVE TABLE bronze_site_usage COMMENT "ingest latest snapshot from api"TBLPROPERTIES ("layer" = "bronze") AS SELECT * FROM cloud_fi...
I realize that the best practice would be to integrate our service with Kafka as a streaming source for Databricks, but given that the service already stores data into TimescaleDB, how can I stream data from TimescaleDB into DBX? Debezium doesn't wor...
When anyone (admins included) click on an alert task in a job run we see the error `No results were found: the query results may no longer be available or you may not have permissions`.Should we be seeing something else or is this a matter of a poor ...
The error message "No results were found: the query results may no longer be available or you may not have permissions" is designed to address a range of potential situations. This includes instances where data might not be accessible due to reasons ...
I have the following command that runs in my databricks notebook.spark.conf.get("spark.databricks.clusterUsageTags.managedResourceGroup")I have wrapped this command into a function (simplified).def get_info(): return spark.conf.get("spark.databri...
Hi,Whenever I try to signup for an account, I keep getting the following message in the first step - "an error has occurred. please try again later" Could you please let me know why this could be? I tried multiple emails and seems to be having same i...
Recently I discovered an issue when creating a PARQUET table that contains a column of type STRUCT with more than 350 string subfields. Such a table can be successfully created via a standard DDL script nevertheless each subsequent attempt to work wi...
HI,When I do GET request to obtain list of job runs using `/api/2.1/jobs/runs/list` there is no `prev_page_token`, `next_page_token` fields in the response despite of having `has_more: True`.
@Hubert-Dudek @Retired_mod could you confirm that behavior is an issue? if yes, can I report it anywhere?Otherwise can I report it as a feature request (to get all runs)? If yes, do you have any service to report it?
The new way of working with Databricks allows users to easily connect their notebook to a SQL serverless warehouse, write code in plain English, and generate output with the press of Space-Ctrl-Shift.
I'm trying to get my workspace setup by connecting to my PostgreSQL database. I am following this article. The instructions under "Create a connection" are failing because my connection requires a database name. However, the database name is not a su...
This is the error.[CANNOT_ESTABLISH_CONNECTION] Cannot establish connection to remote POSTGRESQL database. Please check connection information and credentials e.g. host, port, user and password options. ** If you believe the information is correct, p...
@Sujitha Hi Sujitha, Could you please let us know when we can see the Databricks rewards portal and we hope that the points credited over there will remain the same. Please update on these 2.
@Sujitha Could you please let us know when we can see the Databricks rewards portal. i see its still under construction
I have a set of database-files (.db) which I need to read into my Python Notebook in Databricks. I managed to do this fairly simple up until July when a update in SQLite JDBC library was introduced. Up until now I have read the files in question with...
When the numbers in the table are really big (millions and billions) or really low (e.g. 1e-15), SQLite JDBC may struggle to import the correct values. To combat this, a good idea could be to use customSchema in options to define the schema using Dec...
when we use this cron schedule: 0 58/30 6,7,8,9,10,11,12,13,14,15,16,17 ? * MON,TUE,WED,THU,FRI *so far only the 58th minute will run, but not the 28th minute (30minutes after 58th minute). Is there some kind of bug in the cron scheduler?Reference: h...
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now| User | Count |
|---|---|
| 133 | |
| 128 | |
| 62 | |
| 57 | |
| 42 |