cancel
Showing results for 
Search instead for 
Did you mean: 
Generative AI
Explore discussions on generative artificial intelligence techniques and applications within the Databricks Community. Share ideas, challenges, and breakthroughs in this cutting-edge field.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Sujitha
by Databricks Employee
  • 28924 Views
  • 18 replies
  • 33 kudos

Databricks Announces the Industry’s First Generative AI Engineer Learning Pathway and Certification

Today, we are announcing the industry's first Generative AI Engineer learning pathway and certification to help ensure that data and AI practitioners have the resources to be successful with generative AI. At Databricks, we recognize that generative ...

Screenshot 2024-01-24 at 11.32.01 PM.png
  • 28924 Views
  • 18 replies
  • 33 kudos
Latest Reply
AIChief
New Contributor II
  • 33 kudos

thanks for sharing

  • 33 kudos
17 More Replies
tefrati
by New Contributor III
  • 40 Views
  • 6 replies
  • 1 kudos

getting 500 on embedding model invocation call

I'm getting the following error message "{"error_code": "INTERNAL_ERROR", "message": "The server received an invalid response from an upstream server."}" when making a call to bge-large-en embedding model.

  • 40 Views
  • 6 replies
  • 1 kudos
Latest Reply
nayan_wylde
Honored Contributor III
  • 1 kudos

Yeah your rate limit seems to be good. Can you also check the following points.1. Use the Databricks-specific name (e.g., databricks-bge-large-en), not the Hugging Face model name. Check in Serving → Endpoints.2. Validate Payload Format{ "input": "te...

  • 1 kudos
5 More Replies
Yash01Kumar12
by Visitor
  • 27 Views
  • 2 replies
  • 0 kudos

Agent Bricks Information Extraction

I am facing some problem in Information extraction using PDF. I have done all the necessary steps. 1) I loaded the data in Volume.2) I ran the Use PDF's functionality to create a structure table of the PDFs3) I now have the table with the column name...

Yash01Kumar12_0-1760117409670.png Yash01Kumar12_1-1760117530018.png
  • 27 Views
  • 2 replies
  • 0 kudos
Latest Reply
NandiniN
Databricks Employee
  • 0 kudos

For the issue 2: INVALID_PARAMETER_VALUE: Couldn't find enough valid rows in the selected table. Found 0 rows, minimum required is 1 for agent creation. Since the previous step was failing due to a data type issue (VARIANT), it's highly likely that ...

  • 0 kudos
1 More Replies
Dulce42
by New Contributor II
  • 885 Views
  • 1 replies
  • 0 kudos

Resolved! API GENIE

Hi community!Yesterday I tried extract history chat from my genie spaces but I can't export chats from other users, I have the next error:{'error_code': 'PERMISSION_DENIED', 'message': 'User XXXXXXXX does not own conversation XXXXXXXX', 'details': [{...

  • 885 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Based on current documentation and available resources, exporting chat histories from Genie Spaces is restricted by ownership rules: only the user who owns the conversation can export that specific chat history, regardless of admin permissions or wor...

  • 0 kudos
PiotrM
by New Contributor III
  • 171 Views
  • 7 replies
  • 9 kudos

Resolved! ai_query not affected by AI gateway's rate limits?

Hey, We've been testing the ai_query (Azure Databricks here) on preconfigured model serving endpoints likedatabricks-meta-llama-3-3-70b-instruct and the initial results look nice. I'm trying to limit the number of requests that could be sent to those...

  • 171 Views
  • 7 replies
  • 9 kudos
Latest Reply
PiotrM
New Contributor III
  • 9 kudos

Hey, @BS_THE_ANALYST, before writing that post, I went exactly through the docs you've posted. I wasn't able to find a specific confirmation (or denial) that this function will be affected by the rate limits, which led me to believe that it's worth a...

  • 9 kudos
6 More Replies
qlmahga2
by New Contributor III
  • 496 Views
  • 2 replies
  • 1 kudos

Finally fix Claude Opus

It's been almost 3 months since the announcement of First-party Anthropic Claude Opus 4 on Databricks Mosaic AI Model Serving, but the model is still unavailable.It's listed in the pricing and documentation, but on the Serving endpoints page, it's be...

  • 496 Views
  • 2 replies
  • 1 kudos
Latest Reply
jack_zaldivar
Databricks Employee
  • 1 kudos

Hi @qlmahga2 , it appears that Claude Sonnet 4 and Claude Sonnet 4.5 are both currently available. I'm not sure what it looked like when you asked this question, as I didn't check at that time. However, it looks like they should be available now. Can...

  • 1 kudos
1 More Replies
Nobuhiko
by New Contributor III
  • 1410 Views
  • 8 replies
  • 8 kudos

Resolved! Can I get notebooks used in a e-learning video?

I'm currently watching the videos in the "Generative AI Engineering Pathway."In the "Demo" chapters, it appears that the instructor is explaining based on pre-prepared notebooks (for example, a notebook named "2.1 - Preparing Data for RAG"). Would it...

  • 1410 Views
  • 8 replies
  • 8 kudos
Latest Reply
Nobuhiko
New Contributor III
  • 8 kudos

Hello @Advika !Could you provide details about "the lab materials"?Are they just notebooks, or do they include other materials like raw data files to be processed, environment variables (like "DA"), scripts, etc.?I'm very new to Databricks, so it wou...

  • 8 kudos
7 More Replies
GiriSreerangam
by New Contributor II
  • 148 Views
  • 4 replies
  • 2 kudos

AI Agents - calling custom code and databricks jobs

Hi EveryoneI am building AI Agents; my requirement is to call custom tool logic (which was not possible using a unity catalog function) and databricks jobs. I could not find much of documentation on these scenarios. If someone could share any referen...

  • 148 Views
  • 4 replies
  • 2 kudos
Latest Reply
jamesl
Databricks Employee
  • 2 kudos

Hi @GiriSreerangam , can you share what you are trying to do with the custom tool? You might be able to implement with a custom MCP server. Here are some other resources that may help:  https://github.com/JustTryAI/databricks-mcp-serverhttps://github...

  • 2 kudos
3 More Replies
Phani1
by Valued Contributor II
  • 331 Views
  • 2 replies
  • 2 kudos

Export/Share Genie Space Across DEV, QA, and PROD Environments

Hi Team,What is the procedure for exporting a Genie space across multiple workspace environments such as DEV, QA, and PROD?Can you provide any details around this. Thanks ,Phani

  • 331 Views
  • 2 replies
  • 2 kudos
Latest Reply
spoltier
New Contributor III
  • 2 kudos

@Pilsner When I try to export, I get Error: dataRoom is not user-facing. Trying to download in the workspace browser results in a 400 Bad Request.The bundle documentation for AWS and Azure does not mention genie spaces as existing resources.See the r...

  • 2 kudos
1 More Replies
brahaman
by New Contributor II
  • 353 Views
  • 1 replies
  • 1 kudos

Vectorisation job automatisation and errors

Hey there ! So I'm fairly new to AI and RAG, and at this moment I'm trying to automatically vectorise documents (.pdf, .txt, etc...) each time a new file comes in a volume that I created.For that I created, a job that's triggered each time a new file...

  • 353 Views
  • 1 replies
  • 1 kudos
Latest Reply
mark_ott
Databricks Employee
  • 1 kudos

To address the question about automating and optimizing document vectorization pipelines (PDF, TXT, etc.) like the Databricks unstructured data pipeline with challenges around HuggingFace model downloads and job flexibility, here are insights and alt...

  • 1 kudos
prakashhinduja1
by New Contributor
  • 265 Views
  • 1 replies
  • 1 kudos

Resolved! Prakash Hinduja Switzerland (Swiss) How can I manage spending while optimizing compute resources?

Hi I am Prakash Hinduja Visionary Financial Strategist, born in Amritsar (India) and now lives in Geneva, Switzerland (Swiss) I’m looking for advice on how to better manage costs in Databricks while still keeping performance efficient. If you’ve foun...

  • 265 Views
  • 1 replies
  • 1 kudos
Latest Reply
mark_ott
Databricks Employee
  • 1 kudos

To optimize costs in Databricks while maintaining strong performance, consider a blend of strategic cluster configurations, autoscaling, aggressive job scheduling, and robust monitoring tools. These proven practices are used by leading enterprises in...

  • 1 kudos
chunky35
by New Contributor
  • 816 Views
  • 1 replies
  • 1 kudos

streaming llm response

I am deploying an agent that works good withouth streaming:it is using the following packages:      "mlflow==2.22.1",      "langgraph",      "langchain",      "pydantic==2.8.2",      "langgraph-checkpoint-sqlite",      "databricks-langchain",      "p...

  • 816 Views
  • 1 replies
  • 1 kudos
Latest Reply
mark_ott
Databricks Employee
  • 1 kudos

To implement streaming output for your agent in Databricks and resolve the error "This model does not support predict_stream method.", the key requirement is that your underlying MLflow model must support the predict_stream method. Most l...

  • 1 kudos
HemantvIkani32
by New Contributor II
  • 540 Views
  • 5 replies
  • 1 kudos

Model deprecation issue while serving on Databricks

I am facing this below error while serving codellama model:Exception: Request failed with status 400, {"error_code":"INVALID_PARAMETER_VALUE","message":"The provisioned throughput model Llama 2 7B is deprecated and no longer supported in serving. See...

  • 540 Views
  • 5 replies
  • 1 kudos
Latest Reply
Advika
Databricks Employee
  • 1 kudos

Hello @HemantvIkani32! Did the responses shared above help resolve your concern? If yes, please consider marking the relevant response(s) as the accepted solution.

  • 1 kudos
4 More Replies
brahaman
by New Contributor II
  • 244 Views
  • 1 replies
  • 1 kudos

Agent Serving Not Able To Use Route Optimisation

Hey y'all !I have model that I already serve with `agents.deploy` and works fine.I wanted to try out the Route Optimisation features as it can be found in the Edit tab of a model serving by adding in the creation via the API call (https://docs.databr...

  • 244 Views
  • 1 replies
  • 1 kudos
Latest Reply
sarahbhord
Databricks Employee
  • 1 kudos

Brahaman -  Thanks for reaching out. I hope I am interpreting the question correctly, but route optimization on Databricks serving endpoints can only be enabled at the time of endpoint creation. Update-in-place is not supported, and there is no way t...

  • 1 kudos
dk_g
by New Contributor
  • 467 Views
  • 1 replies
  • 0 kudos

How to utilize clustered gpu for large hf models

Hi,I am using clustered GPU(driver -1GPU and Worker-3GPU), and caching model data into unity catalog but while loading model checkpoint shards its always use driver memory and failed due insufficient memory.How to use complete cluster GPU while loadi...

  • 467 Views
  • 1 replies
  • 0 kudos
Latest Reply
lin-yuan
Databricks Employee
  • 0 kudos

1. Are you using any of the model parallel library, such as FSDP or DeepSpeed? Otherwise, every GPU will load the entire model weights.  2. If yes in 1, Unity Catalog Volumes are exposed on every node at /Volumes/<catalog>/<schema>/<volume>/..., so w...

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now