<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Gen AI governance and compliance in Generative AI</title>
    <link>https://community.databricks.com/t5/generative-ai/gen-ai-governance-and-compliance/m-p/78243#M264</link>
    <description>&lt;P&gt;I have few questions for which I am looking for answers in Databricks context.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;1. Plagiarism and Originality&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp;How do we address the issue of plagiarism, where similar code, solutions, or documentation may be generated for similar prompts globally, potentially leading to intellectual property concerns?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;2. Data Governance and Compliance&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp;What measures do we have in place to ensure compliance with data governance regulations such as GDPR, HIPAA, and other relevant policies, to protect user data and maintain transparency in data handling practices?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;3. Handling Sensitive Data&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp;How do we handle personally identifiable information (PII) and other sensitive data that may be shared through your platform, to prevent unauthorized access, misuse, or exposure?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;Content Moderation and Bias&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp;What steps do we take to detect and prevent the generation of unwanted or inappropriate content, including geo-political or biased viewpoints, in the context of documentation and explanations provided by your LLMs?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;5. Decentralization and Vendor Lock-in&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; How do we mitigate the risk of centralized dependency on large language models (LLMs) and cloud providers, and what alternatives do you offer to ensure users have control over their data and models?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;6. Secure Code Transmission and Sharing&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; What security measures do you have in place to protect proprietary code and sensitive information when transmitted and shared over APIs or the internet to public versions of LLMs&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;Logging and Auditing&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; How do we log and store user prompts and responses, and what mechanisms do you have in place to enable auditing and scrutiny of user activity to detect ethical, unethical, or unlawful practices?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;8. Sensitive Information Storage and Protection&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; What safeguards do you have in place to prevent users from storing sensitive information as prompts within LLMs, and how do you ensure that such information is not inadvertently shared or exposed in the public domain?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;9. Data Utilization and Ownership&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp;&amp;nbsp;How do we address the concern that, according to the terms and conditions of public LLM usage, providers may store all prompts and responses for a duration and utilize them for training purposes, potentially compromising user data ownership and control?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;10. Data Security and Control&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;How do we ensure the security of the data or prompts and responses used for training, testing, and deployment of the trained/tuned model? Is the data in control within the account logged in&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;`11.Distilled/&lt;STRONG&gt;Child Model Updates and Pricing&lt;/STRONG&gt;&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; Once a distilled model is created in our account/project, during training or post-deployment, will the trained model update the global version? Or, post-deployment, do the APIs need to be re-built to point to the custom model? Are there any pricing implications for these custom LLMs for prompt/response &amp;nbsp;when called from an application?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;12. Hosting Custom Models&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;What is the cost impact of hosting this custom model post-training, testing, and deployment?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Thu, 11 Jul 2024 05:11:09 GMT</pubDate>
    <dc:creator>ShankarM</dc:creator>
    <dc:date>2024-07-11T05:11:09Z</dc:date>
    <item>
      <title>Gen AI governance and compliance</title>
      <link>https://community.databricks.com/t5/generative-ai/gen-ai-governance-and-compliance/m-p/78243#M264</link>
      <description>&lt;P&gt;I have few questions for which I am looking for answers in Databricks context.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;1. Plagiarism and Originality&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp;How do we address the issue of plagiarism, where similar code, solutions, or documentation may be generated for similar prompts globally, potentially leading to intellectual property concerns?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;2. Data Governance and Compliance&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp;What measures do we have in place to ensure compliance with data governance regulations such as GDPR, HIPAA, and other relevant policies, to protect user data and maintain transparency in data handling practices?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;3. Handling Sensitive Data&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp;How do we handle personally identifiable information (PII) and other sensitive data that may be shared through your platform, to prevent unauthorized access, misuse, or exposure?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;Content Moderation and Bias&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp;What steps do we take to detect and prevent the generation of unwanted or inappropriate content, including geo-political or biased viewpoints, in the context of documentation and explanations provided by your LLMs?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;5. Decentralization and Vendor Lock-in&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; How do we mitigate the risk of centralized dependency on large language models (LLMs) and cloud providers, and what alternatives do you offer to ensure users have control over their data and models?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;6. Secure Code Transmission and Sharing&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; What security measures do you have in place to protect proprietary code and sensitive information when transmitted and shared over APIs or the internet to public versions of LLMs&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;Logging and Auditing&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; How do we log and store user prompts and responses, and what mechanisms do you have in place to enable auditing and scrutiny of user activity to detect ethical, unethical, or unlawful practices?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;8. Sensitive Information Storage and Protection&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; What safeguards do you have in place to prevent users from storing sensitive information as prompts within LLMs, and how do you ensure that such information is not inadvertently shared or exposed in the public domain?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;9. Data Utilization and Ownership&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp;&amp;nbsp;How do we address the concern that, according to the terms and conditions of public LLM usage, providers may store all prompts and responses for a duration and utilize them for training purposes, potentially compromising user data ownership and control?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;10. Data Security and Control&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;How do we ensure the security of the data or prompts and responses used for training, testing, and deployment of the trained/tuned model? Is the data in control within the account logged in&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;`11.Distilled/&lt;STRONG&gt;Child Model Updates and Pricing&lt;/STRONG&gt;&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; Once a distilled model is created in our account/project, during training or post-deployment, will the trained model update the global version? Or, post-deployment, do the APIs need to be re-built to point to the custom model? Are there any pricing implications for these custom LLMs for prompt/response &amp;nbsp;when called from an application?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;STRONG&gt;12. Hosting Custom Models&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;What is the cost impact of hosting this custom model post-training, testing, and deployment?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 11 Jul 2024 05:11:09 GMT</pubDate>
      <guid>https://community.databricks.com/t5/generative-ai/gen-ai-governance-and-compliance/m-p/78243#M264</guid>
      <dc:creator>ShankarM</dc:creator>
      <dc:date>2024-07-11T05:11:09Z</dc:date>
    </item>
    <item>
      <title>Re: Gen AI governance and compliance</title>
      <link>https://community.databricks.com/t5/generative-ai/gen-ai-governance-and-compliance/m-p/78703#M270</link>
      <description>&lt;P&gt;Thanks&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/9"&gt;@Retired_mod&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am looking for specific solutions on how this can be implemented in databricks for each of the above points. What tools/frameworks/functions can be used. I understand that it will depend on use cases but if you can take one example and guide it will help..&lt;/P&gt;</description>
      <pubDate>Sun, 14 Jul 2024 14:46:01 GMT</pubDate>
      <guid>https://community.databricks.com/t5/generative-ai/gen-ai-governance-and-compliance/m-p/78703#M270</guid>
      <dc:creator>ShankarM</dc:creator>
      <dc:date>2024-07-14T14:46:01Z</dc:date>
    </item>
  </channel>
</rss>

