Background:I've created a small function in a notebook that uses Splunk's splunk-sdk package. The original intention was to call Splunk to execute a search/query, but for the sake of simplicity while testing this issue, the function only prints pr...
I'm new to RANGE_JOIN so this may be completely normal, but I'd like confirmation.Whenever I put a RANGE_JOIN hint in my query SELECT /*+ RANGE_JOIN(pr2, 3600) */
event.FirstIP4Record
FROM SCHEMA_NAME_HERE.dnsrequest event
INNER JOIN SC...
I'm experimenting with liquid clustering and have some questions about compatible types (somewhat similar to Liquid clustering with boolean columns ).Table created as CREATE TABLE IF NOT EXISTS <TABLE>
(
_time DOUBLE
, timestamp TIMESTAMP_NT...
I have some CSV files that I upload to DBFS storage several times a day. From these CSVs, I have created SQL tables: CREATE TABLE IF NOT EXISTS masterdata.lookup_host
USING CSV
OPTIONS (header "true", inferSchema "true")
LOCATION '/mnt/masterdata/...
Another idea (if you need to do small lookups, not bulk transfer) .... what about using Splunk's splunk-sdk to create a notebook function that hits Splunk via REST API?
In my experience with the Splunk add-on, it is typically used to pull Databricks data into Splunk, not to push. If the data sets are small then it could probably push as well, but I think you'd have to write some sort of Splunk map loop to issue I...
I'm not sure if this is related, but I've hit another challenge with TIMESTAMP_NTZ columnsAs soon as I calculate the statistics on a TIMESTAMP_NTZ column in a table, I can't use that column in a WHERE clause date range.This query -- set the variable ...
Running this fills up the statistics for the columns.ANALYZE TABLE <TABLE> COMPUTE STATISTICS FOR COLUMNS timestamp,aid,ContextProcessIdBut I still get the error when I run OPTIMIZE:Unsupported datatype 'TimestampNTZType'
com.databricks.backend.commo...