<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Generate and export dbt documentation from the Workflow dbt task to S3 in Machine Learning</title>
    <link>https://community.databricks.com/t5/machine-learning/generate-and-export-dbt-documentation-from-the-workflow-dbt-task/m-p/32670#M1738</link>
    <description>&lt;P&gt;I'm testing the Databricks Jobs feature with a dbt task and wanted to know if you had any advice for me for managing dbt documentation.&lt;/P&gt;&lt;P&gt;I can use "dbt run" commands to run my models then "dbt docs generate" to generate the documentation. But is it possible to export the generated files to GitHub or to a File System like AWS S3 ?&lt;/P&gt;</description>
    <pubDate>Fri, 02 Sep 2022 13:07:54 GMT</pubDate>
    <dc:creator>Anonymous</dc:creator>
    <dc:date>2022-09-02T13:07:54Z</dc:date>
    <item>
      <title>Generate and export dbt documentation from the Workflow dbt task to S3</title>
      <link>https://community.databricks.com/t5/machine-learning/generate-and-export-dbt-documentation-from-the-workflow-dbt-task/m-p/32670#M1738</link>
      <description>&lt;P&gt;I'm testing the Databricks Jobs feature with a dbt task and wanted to know if you had any advice for me for managing dbt documentation.&lt;/P&gt;&lt;P&gt;I can use "dbt run" commands to run my models then "dbt docs generate" to generate the documentation. But is it possible to export the generated files to GitHub or to a File System like AWS S3 ?&lt;/P&gt;</description>
      <pubDate>Fri, 02 Sep 2022 13:07:54 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/generate-and-export-dbt-documentation-from-the-workflow-dbt-task/m-p/32670#M1738</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2022-09-02T13:07:54Z</dc:date>
    </item>
    <item>
      <title>Re: Generate and export dbt documentation from the Workflow dbt task to S3</title>
      <link>https://community.databricks.com/t5/machine-learning/generate-and-export-dbt-documentation-from-the-workflow-dbt-task/m-p/32672#M1740</link>
      <description>&lt;P&gt;Hi, &lt;/P&gt;&lt;P&gt;Thanks for your answer. Actually, I used this documentation for generating my Databricks jobs but  there is no mention about how to manage the dbt generated documentation. I do not know if that feature is already implemented by Databricks or if there is a work around. &lt;/P&gt;</description>
      <pubDate>Mon, 05 Sep 2022 12:14:53 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/generate-and-export-dbt-documentation-from-the-workflow-dbt-task/m-p/32672#M1740</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2022-09-05T12:14:53Z</dc:date>
    </item>
    <item>
      <title>Re: Generate and export dbt documentation from the Workflow dbt task to S3</title>
      <link>https://community.databricks.com/t5/machine-learning/generate-and-export-dbt-documentation-from-the-workflow-dbt-task/m-p/32673#M1741</link>
      <description>&lt;P&gt;&amp;nbsp;Hi @Kaniz Fatma,&lt;/P&gt;&lt;P&gt;​&lt;/P&gt;&lt;P&gt;The documentation mentions:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Automatic archiving of the artifacts from job runs, including logs, results, manifests, and configuration.&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;​&lt;/P&gt;&lt;P&gt;When a dbt task runs, do the logs, manifests and index.html automatically go back to the attached repo?&lt;/P&gt;&lt;P&gt;​&lt;/P&gt;&lt;P&gt;Is there a way to run slim ci with the dbt task? Can we use pre-commit? It would be good to be able to inspect manifest, capture models that have changed, shallow clone them, test their transforms, and if succeed, run those to prod.​&lt;/P&gt;&lt;P&gt;​&lt;/P&gt;</description>
      <pubDate>Sat, 08 Oct 2022 23:05:16 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/generate-and-export-dbt-documentation-from-the-workflow-dbt-task/m-p/32673#M1741</guid>
      <dc:creator>Jfoxyyc</dc:creator>
      <dc:date>2022-10-08T23:05:16Z</dc:date>
    </item>
    <item>
      <title>Re: Generate and export dbt documentation from the Workflow dbt task to S3</title>
      <link>https://community.databricks.com/t5/machine-learning/generate-and-export-dbt-documentation-from-the-workflow-dbt-task/m-p/32674#M1742</link>
      <description>&lt;P&gt;You can use the jobs api and hit the jobs/runs/get-output/ and look at the dbt_output.artifacts_link to get an http link to download a tar.gz file that has all the artifacts in it. You can then unpack the tar.gz and store those files in adls or s3.&lt;/P&gt;</description>
      <pubDate>Thu, 29 Dec 2022 02:29:46 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/generate-and-export-dbt-documentation-from-the-workflow-dbt-task/m-p/32674#M1742</guid>
      <dc:creator>Jfoxyyc</dc:creator>
      <dc:date>2022-12-29T02:29:46Z</dc:date>
    </item>
    <item>
      <title>Re: Generate and export dbt documentation from the Workflow dbt task to S3</title>
      <link>https://community.databricks.com/t5/machine-learning/generate-and-export-dbt-documentation-from-the-workflow-dbt-task/m-p/72754#M3350</link>
      <description>&lt;P&gt;How can I access these target files from the task itself ? I am trying to use dbt's state modifiers for detecting models that changed and only running models when the source freshness changed. Is there an easy way to store and use these state files in s3/databricks workspace? We are also using Databricks Asset bundles to deploy our workflows and code, so maybe theres a way to use it for this problem ?&lt;/P&gt;</description>
      <pubDate>Wed, 12 Jun 2024 09:49:13 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/generate-and-export-dbt-documentation-from-the-workflow-dbt-task/m-p/72754#M3350</guid>
      <dc:creator>136039</dc:creator>
      <dc:date>2024-06-12T09:49:13Z</dc:date>
    </item>
  </channel>
</rss>

