Tuesday, July 23, 2024

Google parent Alphabet tells workers to be wary of AI chatbots


Alphabet has told its employees not to enter confidential information into Bard, the generative-AI chatbot created and operated by Google, which Alphabet owns.

The company’s warning also extends to other chatbots, such as Microsoft-backed ChatGPT from OpenAI, Reuters reported on Thursday.

Recommended Videos

The AI-powered chatbots have generated huge interest in recent months due to their impressive ability to converse in a human-like way, write essays and reports, and even succeed in academic tests.


  • What is MusicLM? Check out Google’s text-to-music AI

  • ChatGPT: How to use the AI chatbot that’s changing everything

  • ChatGPT creator seeking to eliminate chatbot ‘hallucinations’

But Alphabet has concerns about its workers inadvertently leaking internal data via the tools.

In ongoing work to refine and improve the advanced AI technology, human reviewers may read the conversations that users have with the chatbots, posing a risk to personal privacy and also the potential exposure of trade secrets, the latter of which Alphabet appears to be particularly concerned about.

In addition, the chatbots are partly trained using users’ text exchanges, so with certain prompts the tool could potentially repeat confidential information that it receives in those conversations to members of the public.

Like ChatGPT, Bard is now freely available for anyone to try. On its webpage, it warns users: “Please do not include information that can be used to identify you or others in your Bard conversations.”

It adds that Google collects “Bard conversations, related product usage information, info about your location, and your feedback,” and uses the data to improve Google products and services that include Bard.

Google says it stores Bard activity for up to 18 months, though a user can change this to 3 or 36 months in their Google account.

It adds that as a privacy measure, Bard conversations are disconnected from a Google account before a human reviewer sees them.

Reuters said that while Alphabet’s warning has been in place for a while, it recently expanded it, telling its workers to avoid using precise computer code generated by chatbots. The company told the news outlet that Bard can sometimes make “undesired code suggestions,” though the current iteration of the tool is still considered to be a viable programming aid.

Alphabet isn’t the only company to warn its employees about the privacy and security risks linked to using the chatbots. Samsung recently issued a similar instruction to its workers after a number of them fed sensitive semiconductor-related data into ChatGPT, and Apple and Amazon, among others, have reportedly also enacted a similar internal policy.

Read more

More News