Privatisation of knowledge: why Google Bard and ChatGPT are a problem for us
In July, Google changed its privacy policy to allow its in-house AI Bard to use the entire public Internet for training. That’s a problem.
The Internet thrives on the exchange of information. People share their knowledge on blogs, social media, YouTube, forums and Wikipedia, just to name a few. And if others want to access this information, they have to search for it on the respective websites.
But what if that’s no longer necessary? What if searching and browsing becomes obsolete through ChatGPT or Google Bard? After all, these tools simply take information and make it available to you directly. On the one hand, this has its advantages. You save time and effort. But in the long run, you’ll pay a high price for your comfort.
After all, what happens when no one visits Wikipedia any more because users have become accustomed to AI tools? Users, donors and volunteer authors on Wikipedia will disappear. Information will become outdated, new pages won’t be added. And available, public knowledge becomes privatised.
Little incentive to share new knowledge
Providers such as Open AI and Google are already taking advantage of the general knowledge available on the Internet and building paid services. They profit from what the general public has created, often paid for by the public, universities or foundations such as Wikipedia.
When we extrapolate the success of AI services into the future, many issues come to light. One of them being that less and less information could be published due to a lack of visibility for creators. Whether it’s recognition or money through advertising or subscriptions – it’ll all evaporate. And then? The creation of new information will come to a standstill. The system breaks down. The knowledge that until now has been put online publicly will be solely held by corporations. And this knowledge can be dosed, concealed or adjusted as desired.
How is an AI different from a human who knows the entire Internet?
But it can’t be that bad, right? Don’t these new tools just do what humans do but more efficiently? Only partly. A human being who wants to educate themselves on a topic and reads and looks at everything publicly available to them isn’t comparable to an AI. The difference is, this person will never be able to aggregate all the information. And if they could, then only one person would have access to the knowledge – not everyone who uses a tool. A tool owned by a corporation.
A topic that affects us all
The privatisation of knowledge is a complex issue that affects us all. It’s time to think about what this means for our future. The difference between reading publicly available and privatised information is more than just a matter of access. It’s a central question moving into the future.
How do you see this development?
Cool: Creating interfaces between the real world and the world of pure information. Not cool: Driving by car to the mall to shop. My life happens online, the information age is where I feel at home.