/* ---- Google Analytics Code Below */

Sunday, February 12, 2023

The Problem with the Privacy of ChatGPT Data

 Very key issue, when we converse with any chatbot, they have learned something about our needs. To predict your next question.    As Chatbots gets more sophisticated, they can get more specific and valuable data.    Are we ready to do that?   Is a warning sign-off enough?  

ChatGPT is a data privacy nightmare, and we ought to be concerned  in ArsTechnica.  

ChatGPT's extensive language model is fueled by our personal data.


ChatGPT has taken the world by storm. Within two months of its release it reached 100 million active users, making it the fastest-growing consumer application ever launched. Users are attracted to the tool’s advanced capabilities—and concerned by its potential to cause disruption in various sectors.

A much less discussed implication is the privacy risks ChatGPT poses to each and every one of us. Just yesterday, Google unveiled its own conversational AI called Bard, and others will surely follow. Technology companies working on AI have well and truly entered an arms race.

The problem is, it’s fueled by our personal data.

300 billion words. How many are yours?

ChatGPT is underpinned by a large language model that requires massive amounts of data to function and improve. The more data the model is trained on, the better it gets at detecting patterns, anticipating what will come next, and generating plausible text.

OpenAI, the company behind ChatGPT, fed the tool some 300 billion words systematically scraped from the Internet: books, articles, websites, and posts—including personal information obtained without consent.

If you’ve ever written a blog post or product review, or commented on an article online, there’s a good chance this information was consumed by ChatGPT.

So why is that an issue?

The data collection used to train ChatGPT is problematic for several reasons.

First, none of us were asked whether OpenAI could use our data. This is a clear violation of privacy, especially when data is sensitive and can be used to identify us, our family members, or our location.  ... '   ... 

No comments: