<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: what are the four-phase of transformation where catalyst  transformation is used in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/what-are-the-four-phase-of-transformation-where-catalyst/m-p/19602#M13158</link>
    <description>&lt;P&gt;1. Analysis&lt;/P&gt;&lt;P&gt;The first phase of Spark SQL optimization is the analysis. Spark SQL starts with a relationship to be processed that can be in two ways. A serious form from an AST (abstract syntax tree) returned by an SQL parser, and on the other hand from a DataFrame object of the Spark SQL API.&lt;/P&gt;&lt;P&gt;2.&amp;nbsp;Logic Optimization Plan&lt;/P&gt;&lt;P&gt;The second phase is the logical optimization plan. In this phase, rule-based optimization is applied to the logical plan. It is possible to easily add new rules.&lt;/P&gt;&lt;P&gt;3. Physical plan&lt;/P&gt;&lt;P&gt;In the physical plan phase, Spark SQL takes the logical plan and generates one or more physical plans using the physical operators that match the Spark execution engine. The plan to be executed is selected using the cost-based model (comparison between model costs).&lt;/P&gt;&lt;P&gt;4. Code generation&lt;/P&gt;&lt;P&gt;Code generation is the final phase of optimizing Spark SQL. To run on each machine, it is necessary to generate Java code bytecode.&lt;/P&gt;</description>
    <pubDate>Fri, 25 Jun 2021 16:10:26 GMT</pubDate>
    <dc:creator>User16826994223</dc:creator>
    <dc:date>2021-06-25T16:10:26Z</dc:date>
    <item>
      <title>what are the four-phase of transformation where catalyst  transformation is used</title>
      <link>https://community.databricks.com/t5/data-engineering/what-are-the-four-phase-of-transformation-where-catalyst/m-p/19601#M13157</link>
      <description />
      <pubDate>Fri, 25 Jun 2021 16:10:05 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/what-are-the-four-phase-of-transformation-where-catalyst/m-p/19601#M13157</guid>
      <dc:creator>User16826994223</dc:creator>
      <dc:date>2021-06-25T16:10:05Z</dc:date>
    </item>
    <item>
      <title>Re: what are the four-phase of transformation where catalyst  transformation is used</title>
      <link>https://community.databricks.com/t5/data-engineering/what-are-the-four-phase-of-transformation-where-catalyst/m-p/19602#M13158</link>
      <description>&lt;P&gt;1. Analysis&lt;/P&gt;&lt;P&gt;The first phase of Spark SQL optimization is the analysis. Spark SQL starts with a relationship to be processed that can be in two ways. A serious form from an AST (abstract syntax tree) returned by an SQL parser, and on the other hand from a DataFrame object of the Spark SQL API.&lt;/P&gt;&lt;P&gt;2.&amp;nbsp;Logic Optimization Plan&lt;/P&gt;&lt;P&gt;The second phase is the logical optimization plan. In this phase, rule-based optimization is applied to the logical plan. It is possible to easily add new rules.&lt;/P&gt;&lt;P&gt;3. Physical plan&lt;/P&gt;&lt;P&gt;In the physical plan phase, Spark SQL takes the logical plan and generates one or more physical plans using the physical operators that match the Spark execution engine. The plan to be executed is selected using the cost-based model (comparison between model costs).&lt;/P&gt;&lt;P&gt;4. Code generation&lt;/P&gt;&lt;P&gt;Code generation is the final phase of optimizing Spark SQL. To run on each machine, it is necessary to generate Java code bytecode.&lt;/P&gt;</description>
      <pubDate>Fri, 25 Jun 2021 16:10:26 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/what-are-the-four-phase-of-transformation-where-catalyst/m-p/19602#M13158</guid>
      <dc:creator>User16826994223</dc:creator>
      <dc:date>2021-06-25T16:10:26Z</dc:date>
    </item>
  </channel>
</rss>

