DeepSeek is Sharing!

DeepSeek is Sharing!

Author's Note: I summarized the concern based on limited research, and shared it. Further research (Thanks, Jessie!) has led me to rewrite and lose some of the snark (it's really not my thing anyway; I should stick to direct; I'm better at that). Below is the rewrite.

===

Recent findings by security researchers have uncovered a serious security risk with DeepSeek R1, a widely used AI chatbot. Researchers found that the chatbot is transmitting user data directly to?China Mobile’s CMPassport.com, a state-owned telecom company under U.S. sanctions. This is not a minor oversight, it’s a clear security concern that could have serious implications for user privacy and data protection.

The Evidence: Data Leaks and Tracking

The investigation revealed?encrypted code?embedded in DeepSeek’s browser-based version, designed to send user data back to China Mobile. But that’s not all, the chatbot is also?tracking users?by creating digital fingerprints, allowing it to monitor online activity across multiple sites. This means it’s not just storing what you type, but also mapping out your online behavior. (Like many online tools today)

Example: How Digital Fingerprinting Works

For instance, let’s say a user interacts with DeepSeek on their work laptop. The chatbot records unique identifiers such as the user’s IP address, browser settings, device information, and behavioral patterns (e.g., typing speed and cursor movements). These identifiers create a?digital fingerprint, allowing DeepSeek to recognize the user even if they access the chatbot from a different site or browser session. Over time, this fingerprint enables tracking across multiple platforms, meaning that the chatbot, or any entity accessing its logs, can build a detailed profile of the user’s online behavior. If this data is being transmitted to?CMPassport.com, it raises significant concerns about who has access to it and how it might be used.


Why This Matters

The rapid adoption of AI tools like DeepSeek means that businesses, developers, and even government agencies may unknowingly expose sensitive data. This discovery raises major concerns about data security, potential foreign surveillance, and the broader implications of using AI models linked to entities with known cybersecurity risks.

Implementing this model

If you or your organization are using DeepSeek, consider a few implementation options:

  • Implement Network Monitoring?– Set up network traffic analysis tools to detect unauthorized data transmissions to foreign or unknown servers.
  • Restrict AI Tool Access?– Instead of outright banning tools, limit access to non-sensitive tasks or isolate them in controlled environments.
  • Use Sandboxed Environments?– Deploy AI tools in virtualized or air-gapped environments to prevent them from interacting with sensitive internal networks.
  • Advocate for Transparent AI?– Push for AI providers to disclose where data is stored, how it’s processed, and whether third parties have access.
  • Explore Alternative AI Models—Consider self-hosted or open-source AI solutions that allow you to have full control over data security.

AI is a powerful tool, but it must be deployed responsibly. Understanding where your data is going and who has access to it is critical in today’s digital landscape. Security and privacy should never be an afterthought.

</dave>

Jesse Houston

Co-Founder, Game Director at Critical Path Games

1 周

Just read the article. The fine tuned model doesn’t send data back. Their website sends tracking data. Frankly, total nothing burger here. Real news would be if the open weight models send data back somehow hijacking Llama.cpp or Ollama or something Edit: the article https://www.feroot.com/news/the-independent-feroot-security-uncovers-deepseeks-hidden-code-sending-user-data-to-china/#:~:text=Today's%20investigation%20by%20cybersecurity%20firm,online%20registry%2C%20CMPassport.com.

Harry Allen

Innovating with Insight: Strategic Thinker & Storyteller

1 周

It's a shame this needed to be said... But it needed to be said... Shadow AI use in your orgs is taking the legs right out from under your security guardrails. Edit following the Edit: Thanks for the update and for pinging me. The facts remain. LLMs may say they won't use your data but we've been around long enough to know between Cookies and Privacy agreements it's still impossible to contain and protect your data. It only takes ONE slip to leak information that violates regulations and can cost you big time. Thanks for adding that section about safe Implementation. It's been interesting using AI models on my personal device while connecting to public wifi. Many orgs are already outright blocking access to the tools (not uniformly either). I (personally) think self-hosted and/or open-source is likely the way to go. It's definitely something that we at Veritone espouse as a solid option with our deployments (though we offer both cloud and on-prem). Great topic Dave thanks for the insights!

要查看或添加评论,请登录

Dave B.的更多文章