<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>#PrivacyByDesign on Home</title>
    <link>https://yakinin.com/en/tags/%23privacybydesign/</link>
    <description>Recent content in #PrivacyByDesign on Home</description>
    <generator>Hugo -- 0.148.2</generator>
    <language>en</language>
    <lastBuildDate>Fri, 22 Aug 2025 08:23:06 +0000</lastBuildDate>
    <atom:link href="https://yakinin.com/en/tags/%23privacybydesign/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Grok&#39;s Public Chats: A Predictable AI Privacy Failure</title>
      <link>https://yakinin.com/en/posts/20250822-grok-ai-privacy-failure/</link>
      <pubDate>Fri, 22 Aug 2025 08:23:06 +0000</pubDate>
      <guid>https://yakinin.com/en/posts/20250822-grok-ai-privacy-failure/</guid>
      <description>&lt;p&gt;It’s a classic story at this point. We saw it recently with OpenAI’s ChatGPT, and now it’s Grok’s turn. Elon Musk’s xAI has inadvertently published hundreds of thousands of its users&amp;rsquo; private conversations, making them fully searchable on Google. This wasn&amp;rsquo;t a sophisticated hack; it was a fundamental product design flaw.&lt;/p&gt;
&lt;h2 id=&#34;the-feature-that-became-a-bug&#34;&gt;The Feature That Became a Bug&lt;/h2&gt;
&lt;p&gt;The mechanism was simple and naive. When a Grok user hit the &amp;ldquo;share&amp;rdquo; button to send a conversation to a colleague or friend, the system generated a unique URL. However, instead of being a private link, this URL was made public and available for search engines to index. In effect, &amp;ldquo;sharing&amp;rdquo; meant &amp;ldquo;publishing to the open web&amp;rdquo; without any warning or disclaimer.&lt;/p&gt;</description>
    </item>
  </channel>
</rss>
