cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Lakebridge Analyzer stopped working.

Aloknath_Ganage
New Contributor II

Hi There,

I was using the lakebridge Analyzer and tranpiler for the last 2 months and it was working fine and was providing the expected output. But from the last 2-3 days when I'm running the Analyzer command for any of the dialect it is generating analyzer_output.temp file along with the analyzer_output.xlsx.

The xlsx file doesn't containing any data with missing the required sheets in it but the temp file has all the required data, and it is failing to merge.

What could be the problem?

Please find the attached SS for the reference.

 

2 REPLIES 2

Ashwin_DSA
Databricks Employee
Databricks Employee

Hi @Aloknath_Ganage  - Thanks for the details and screenshots.

On doing some research, I believe the Analyzer itself is running correctly... thatโ€™s why you see the analyzer_output.tmp file with all the detailed metrics. The issue is likely in the second step, where Lakebridge takes that temp file and merges it into the Excel workbook. That merge step is currently failing, which is why the .xlsx file only shows the empty โ€œSummary / SQL Programs / SQL Script Categories / UNKNOWN SQL Categoryโ€ tabs with zeros, even though the .tmp file has data.

This is almost always caused by a local Lakebridge/Analyzer installation issue (version mismatch or corruption of the Excel writer/template), not by your input code or the source dialect.

To fix it, please:

  • Make sure the Excel report isnโ€™t open when you run the Analyzer and that the output folder is writable.
  • Reinstall Lakebridge on that machine:
    bash
    databricks labs uninstall lakebridge
    databricks labs install lakebridge
    databricks labs lakebridge install-transpile   # if youโ€™re also using the transpiler
  • Re-run the Analyzer once with debug logging and a small test folder:

    bash

    databricks labs lakebridge analyze \ --source-directory <path> \ --report-file <path>\lakebridge_analysis.xlsx \ --source-tech <SOURCE_TECH> \ --debug
     
    After this run, the Excel should hopefully be populated again. If it still comes out empty, please share the tail of the --debug output and the new .tmp + .xlsx files so I can investigate or talk to our Lakebridge team.
     

    If this answer resolves your question, could you mark it as โ€œAccept as Solutionโ€? That helps other users quickly find the correct fix.

    Regards,  
Regards,
Ashwin | Delivery Solution Architect @ Databricks
Helping you build and scale the Data Intelligence Platform.
***Opinions are my own***

SteveOstrowski
Databricks Employee
Databricks Employee

Hi @Aloknath_Ganage,

The symptoms you describe (the .tmp file contains all the data but the .xlsx is empty with only placeholder sheet tabs) point to the Excel report merge/generation step failing after the core analysis completes successfully.

Here are a few things to check and try:

1. CHECK YOUR LAKEBRIDGE VERSION

Run this to see which version you are on:

databricks labs installed

If you were recently auto-updated to v0.12.1 or v0.12.2, the analyzer filepath handling changed in v0.12.1 (released Feb 12, 2026). That release simplified how --report-file and --source-directory paths are resolved and removed the old "timestamped directory" behavior. If your workflow or wrapper scripts relied on the previous path logic, that could cause the Excel writer to fail silently while the .tmp data is still generated correctly.

2. REINSTALL LAKEBRIDGE

A clean reinstall resolves most merge/Excel generation issues caused by version mismatches or corrupted local state:

databricks labs uninstall lakebridge
databricks labs install lakebridge

If you also use the transpiler:

databricks labs lakebridge install-transpile

3. MAKE SURE NOTHING IS LOCKING THE OUTPUT FILE

Before running the analyzer, confirm that:
- The target .xlsx file is not open in Excel or another process
- The output folder is writable by the current user
- There is no leftover .xlsx from a previous run that might be read-only

4. RE-RUN WITH DEBUG LOGGING

After reinstalling, test with a small input folder and the --debug flag to capture the full trace:

databricks labs lakebridge analyze \
--source-directory /path/to/small-test-folder \
--report-file /path/to/output/lakebridge_analysis.xlsx \
--source-tech <YOUR_SOURCE_TECH> \
--debug

The debug output will show exactly where the merge step fails if it happens again.

5. FILE A GITHUB ISSUE IF THE PROBLEM PERSISTS

Lakebridge is a Databricks Labs project and the team tracks bugs on GitHub. If the reinstall does not fix it, please open an issue at:

https://github.com/databrickslabs/lakebridge/issues

Include the tail end of the --debug output along with the .tmp and .xlsx files. The team is responsive and can look into any regression in the Excel writer.

ADDITIONAL CONTEXT

You can also check the release notes for v0.12.1 and v0.12.2 to see if any of the analyzer changes relate to your setup:

https://github.com/databrickslabs/lakebridge/releases

* This reply used an agent system I built to research and draft this response based on the wide set of documentation I have available and previous memory. I personally review the draft for any obvious issues and for monitoring system reliability and update it when I detect any drift, but there is still a small chance that something is inaccurate, especially if you are experimenting with brand new features.

If this answer resolves your question, could you mark it as "Accept as Solution"? That helps other users quickly find the correct fix.