Alphabet Inc., the parent company of Google, is cautioning its employees about using chatbots, including ChatGPT & its own creation, Bard. This warning comes as the company expands the reach of its chatbot program globally. Let’s delve into the details of this development and understand the underlying concerns.
Also Read: Samsung Bans Employees From Using Generative AI Due to Security Concerns
Safeguarding Confidential Information
According to sources familiar with the matter, Alphabet has instructed its staff to refrain from entering confidential materials into AI chatbots, including Bard. This directive aligns with the company’s longstanding policy on information protection. While the chatbots, such as Bard and ChatGPT, are designed to engage in conversations with users and provide responses, there is a potential risk of data leakage as the AI models can reproduce absorbed training data.
Also Read: AI Is Stealing Your Data – Say Experts
Cautions for Engineers and Programmers
In addition to advising against entering confidential information, Alphabet has alerted its engineers to avoid directly using chatbots’ computer code. While Bard may offer suggestions, it is important for programmers to exercise caution. Google aims to maintain transparency by acknowledging the limitations of its technology and ensuring it does not cause unintended consequences.
Also Read: Apple Follows its Rival Samsung, Bans ChatGPT Over Privacy Fears
A Competitive Landscape and Business Implications
Google’s wariness about chatbot usage stems from its competition with ChatGPT, backed by OpenAI and Microsoft Corp. The stakes are high, with billions of dollars in investments, potential advertising revenues, and cloud revenue riding on the success of these AI programs. Google’s precautions are an industry-wide trend, with other companies like Samsung, Amazon.com, and Deutsche Bank also implementing guardrails for AI chatbots.
Also Read: Microsoft and OpenAI Clash Over AI Integration
Employees’ Use of AI Tools and Security Standards
A survey among professionals from top US-based companies revealed that approximately 43% of respondents used AI tools like ChatGPT, often without informing their superiors. To mitigate potential risks, companies worldwide, including Apple (unconfirmed), have adopted security standards to warn employees about using publicly-available chat programs.
Privacy Concerns and Regulatory Dialogue
Google has engaged in detailed discussions with Ireland’s Data Protection Commission to address concerns related to privacy and comply with regulatory requirements. A recent Politico report stated that the launch of Bard in the European Union was postponed pending additional information on its privacy implications. Google’s updated privacy notice advises users not to include confidential or sensitive information in their Bard conversations.
Also Read: Europe’s Data Protection Board Forms ChatGPT Privacy Task Force
Mitigating Risks with Innovative Solutions
Companies are actively developing software solutions to address these concerns. For instance, Cloudflare, a leading provider of cybersecurity &cloud services, lets businesses tag and restrict certain data from transmitting externally. Google and Microsoft also offer conversational tools to business customers, ensuring data privacy by not incorporating them into public AI models. Although the default setting in Bard and ChatGPT saves users’ conversation history, users can delete it.
Our Say
Google’s warning to its staff regarding chatbot usage, including Bard, reflects the company’s commitment to data privacy and security. As AI technologies continue to evolve, it is crucial for organizations to implement safeguards and promote responsible usage. The dynamic landscape of AI chatbots demands a delicate balance between innovation and risk mitigation. By addressing these concerns, companies like Google are working towards a future where AI technologies can be harnessed safely & ethically.